{"id":189,"date":"2021-07-26T16:43:53","date_gmt":"2021-07-26T16:43:53","guid":{"rendered":"https:\/\/groups.cs.umass.edu\/kdl\/?page_id=189"},"modified":"2022-02-02T19:14:01","modified_gmt":"2022-02-02T19:14:01","slug":"publications","status":"publish","type":"page","link":"https:\/\/groups.cs.umass.edu\/kdl\/publications\/","title":{"rendered":"Publications"},"content":{"rendered":"\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-flow wp-block-group-is-layout-flow\">\n<div class=\"wp-block-buttons is-content-justification-left is-layout-flex wp-block-buttons-is-layout-flex\">\n<div class=\"wp-block-button is-style-outline is-style-outline--1\"><a class=\"wp-block-button__link has-black-color has-text-color\" href=\"https:\/\/groups.cs.umass.edu\/kdl\/publications-search\/\" target=\"_blank\" rel=\"https:\/\/groups.cs.umass.edu\/kdl\/publications-search\/ noopener\">Publications Search<\/a><\/div>\n<\/div>\n<\/div><\/div>\n\n\n\n<div class=\"teachpress_pub_list\"><form name=\"tppublistform\" method=\"get\"><a name=\"tppubs\" id=\"tppubs\"><\/a><\/form><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">127 entries<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 of 3 <a href=\"https:\/\/groups.cs.umass.edu\/kdl\/publications\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"next page\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/groups.cs.umass.edu\/kdl\/publications\/?limit=3&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"last page\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><div class=\"teachpress_publication_list\"><h3 class=\"tp_h3\" id=\"tp_h3_2021\">2021<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Terrance E. Boult, Przemyslaw A. Grabowicz, D. S. Prijatelj, R. Stern, L. Holder, J. Alspector, M. Jafarzadeh, T. Ahmad, A. R. Dhamija, C. Li, S. Cruz, A. Shrivastava, C. Vondrick, W. J. Scheirer<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('841','tp_abstract')\" style=\"cursor:pointer;\">A Unifying Framework for Formal Theories of Novelty:Framework, Examples and Discussion<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">AAAI'21 SMPT, <\/span><span class=\"tp_pub_additional_year\">2021<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 23318422<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_841\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('841','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_841\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('841','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_841\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('841','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_841\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{Boult2020a,<br \/>\r\ntitle = {A Unifying Framework for Formal Theories of Novelty:Framework, Examples and Discussion},<br \/>\r\nauthor = {Terrance E. Boult and Przemyslaw A. Grabowicz and D. S. Prijatelj and R. Stern and L. Holder and J. Alspector and M. Jafarzadeh and T. Ahmad and A. R. Dhamija and C. Li and S. Cruz and A. Shrivastava and C. Vondrick and W. J. Scheirer},<br \/>\r\nurl = {http:\/\/arxiv.org\/abs\/2012.04226},<br \/>\r\nissn = {23318422},<br \/>\r\nyear  = {2021},<br \/>\r\ndate = {2021-12-01},<br \/>\r\nbooktitle = {AAAI'21 SMPT},<br \/>\r\nabstract = {Managing inputs that are novel, unknown, or out-of-distribution is critical as an agent moves from the lab to the open world. Novelty-related problems include being tolerant to novel perturbations of the normal input, detecting when the input includes novel items, and adapting to novel inputs. While significant research has been undertaken in these areas, a noticeable gap exists in the lack of a formalized definition of novelty that transcends problem domains. As a team of researchers spanning multiple research groups and different domains, we have seen, first hand, the difficulties that arise from ill-specified novelty problems, as well as inconsistent definitions and terminology. Therefore, we present the first unified framework for formal theories of novelty and use the framework to formally define a family of novelty types. Our framework can be applied across a wide range of domains, from symbolic AI to reinforcement learning, and beyond to open world image recognition. Thus, it can be used to help kick-start new research efforts and accelerate ongoing work on these important novelty-related problems. This extended version of our AAAI 2021 paper included more details and examples in multiple domains.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('841','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_841\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Managing inputs that are novel, unknown, or out-of-distribution is critical as an agent moves from the lab to the open world. Novelty-related problems include being tolerant to novel perturbations of the normal input, detecting when the input includes novel items, and adapting to novel inputs. While significant research has been undertaken in these areas, a noticeable gap exists in the lack of a formalized definition of novelty that transcends problem domains. As a team of researchers spanning multiple research groups and different domains, we have seen, first hand, the difficulties that arise from ill-specified novelty problems, as well as inconsistent definitions and terminology. Therefore, we present the first unified framework for formal theories of novelty and use the framework to formally define a family of novelty types. Our framework can be applied across a wide range of domains, from symbolic AI to reinforcement learning, and beyond to open world image recognition. Thus, it can be used to help kick-start new research efforts and accelerate ongoing work on these important novelty-related problems. This extended version of our AAAI 2021 paper included more details and examples in multiple domains.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('841','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_841\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/arxiv.org\/abs\/2012.04226\" title=\"http:\/\/arxiv.org\/abs\/2012.04226\" target=\"_blank\">http:\/\/arxiv.org\/abs\/2012.04226<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('841','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Aarshee Mishra, Przemyslaw A. Grabowicz, Nicholas Perello<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('844','tp_abstract')\" style=\"cursor:pointer;\">Towards Fair and Explainable Supervised Learning<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">ICML Workshop on Socially Responsible Machine Learning, <\/span><span class=\"tp_pub_additional_year\">2021<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_844\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('844','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_844\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('844','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_844\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('844','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_844\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{Mishra2021,<br \/>\r\ntitle = {Towards Fair and Explainable Supervised Learning},<br \/>\r\nauthor = {Aarshee Mishra and Przemyslaw A. Grabowicz and Nicholas Perello},<br \/>\r\nurl = {https:\/\/drive.google.com\/file\/d\/1z24hITF0Xrlc_IX_rOZVZ2aigOj1hxhD\/view?usp=sharing},<br \/>\r\nyear  = {2021},<br \/>\r\ndate = {2021-01-01},<br \/>\r\nbooktitle = {ICML Workshop on Socially Responsible Machine Learning},<br \/>\r\nabstract = {Algorithms that aid human decision-making may inadvertently discriminate against certain protected groups. We formalize direct discrimination as a direct causal effect of the protected attributes on the decisions, while textitinduced indirect discrimination as a change in the influence of non-protected features associated with the protected attributes. The measurements of average treatment effect (ATE) and SHapley Additive exPlanations (SHAP) reveal that state-of-the-art fair learning methods can inadvertently induce indirect discrimination in synthetic and real-world datasets. To inhibit discrimination in algorithmic systems, we propose to nullify the influence of the protected attribute on the output of the system, while preserving the influence of remaining features. To achieve this objective, we introduce a risk minimization method which optimizes for the proposed fairness objective. We show that the method leverages model accuracy and disparity measures.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('844','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_844\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Algorithms that aid human decision-making may inadvertently discriminate against certain protected groups. We formalize direct discrimination as a direct causal effect of the protected attributes on the decisions, while textitinduced indirect discrimination as a change in the influence of non-protected features associated with the protected attributes. The measurements of average treatment effect (ATE) and SHapley Additive exPlanations (SHAP) reveal that state-of-the-art fair learning methods can inadvertently induce indirect discrimination in synthetic and real-world datasets. To inhibit discrimination in algorithmic systems, we propose to nullify the influence of the protected attribute on the output of the system, while preserving the influence of remaining features. To achieve this objective, we introduce a risk minimization method which optimizes for the proposed fairness objective. We show that the method leverages model accuracy and disparity measures.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('844','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_844\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/drive.google.com\/file\/d\/1z24hITF0Xrlc_IX_rOZVZ2aigOj1hxhD\/view?usp=sharing\" title=\"https:\/\/drive.google.com\/file\/d\/1z24hITF0Xrlc_IX_rOZVZ2aigOj1hxhD\/view?usp=shari[...]\" target=\"_blank\">https:\/\/drive.google.com\/file\/d\/1z24hITF0Xrlc_IX_rOZVZ2aigOj1hxhD\/view?usp=shari[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('844','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Amanda M Gentzel, Purva Pruthi, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('719','tp_abstract')\" style=\"cursor:pointer;\">How and Why to Use Experimental Data to Evaluate Methods for Observational Causal Inference<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">International Conference on Machine Learning, <\/span><span class=\"tp_pub_additional_pages\">pp. 3660\u20133671, <\/span><span class=\"tp_pub_additional_organization\">PMLR <\/span><span class=\"tp_pub_additional_year\">2021<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_719\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('719','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_719\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('719','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_719\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('719','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_719\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{gentzel2021and,<br \/>\r\ntitle = {How and Why to Use Experimental Data to Evaluate Methods for Observational Causal Inference},<br \/>\r\nauthor = {Amanda M Gentzel and Purva Pruthi and David Jensen},<br \/>\r\nurl = {http:\/\/proceedings.mlr.press\/v139\/gentzel21a\/gentzel21a.pdf},<br \/>\r\nyear  = {2021},<br \/>\r\ndate = {2021-01-01},<br \/>\r\nbooktitle = {International Conference on Machine Learning},<br \/>\r\npages = {3660--3671},<br \/>\r\norganization = {PMLR},<br \/>\r\nabstract = {Methods that infer causal dependence from observational data are central to many areas of science, including medicine, economics, and the social sciences. A variety of theoretical properties of these methods have been proven, but empirical evaluation remains a challenge, largely due to the lack of observational data sets for which treatment effect is known. We describe and analyze observational sampling from randomized controlled trials (OSRCT), a method for evaluating causal inference methods using data from randomized controlled trials (RCTs). This method can be used to create constructed observational data sets with corresponding unbiased estimates of treatment effect, substantially increasing the number of data sets available for evaluating causal inference methods. We show that, in expectation, OSRCT creates data sets that are equivalent to those produced by randomly sampling from empirical data sets in which all potential outcomes are available. We then perform a large-scale evaluation of seven causal inference methods over 37 data sets, drawn from RCTs, as well as simulators, real-world computational systems, and observational data sets augmented with a synthetic response variable. We find notable performance differences when comparing across data from different sources, demonstrating the importance of using data from a variety of sources when evaluating any causal inference method.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('719','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_719\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Methods that infer causal dependence from observational data are central to many areas of science, including medicine, economics, and the social sciences. A variety of theoretical properties of these methods have been proven, but empirical evaluation remains a challenge, largely due to the lack of observational data sets for which treatment effect is known. We describe and analyze observational sampling from randomized controlled trials (OSRCT), a method for evaluating causal inference methods using data from randomized controlled trials (RCTs). This method can be used to create constructed observational data sets with corresponding unbiased estimates of treatment effect, substantially increasing the number of data sets available for evaluating causal inference methods. We show that, in expectation, OSRCT creates data sets that are equivalent to those produced by randomly sampling from empirical data sets in which all potential outcomes are available. We then perform a large-scale evaluation of seven causal inference methods over 37 data sets, drawn from RCTs, as well as simulators, real-world computational systems, and observational data sets augmented with a synthetic response variable. We find notable performance differences when comparing across data from different sources, demonstrating the importance of using data from a variety of sources when evaluating any causal inference method.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('719','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_719\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/proceedings.mlr.press\/v139\/gentzel21a\/gentzel21a.pdf\" title=\"http:\/\/proceedings.mlr.press\/v139\/gentzel21a\/gentzel21a.pdf\" target=\"_blank\">http:\/\/proceedings.mlr.press\/v139\/gentzel21a\/gentzel21a.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('719','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('736','tp_abstract')\" style=\"cursor:pointer;\">Improving Causal Inference by Increasing Model Expressiveness<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the AAAI Conference on Artificial Intelligence, <\/span><span class=\"tp_pub_additional_pages\">pp. 15053\u201315057, <\/span><span class=\"tp_pub_additional_year\">2021<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_736\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('736','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_736\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('736','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_736\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('736','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_736\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{jensen2021improving,<br \/>\r\ntitle = {Improving Causal Inference by Increasing Model Expressiveness},<br \/>\r\nauthor = {David Jensen},<br \/>\r\nurl = {https:\/\/www.aaai.org\/AAAI21Papers\/SMT-427.JensenD.pdf},<br \/>\r\nyear  = {2021},<br \/>\r\ndate = {2021-01-01},<br \/>\r\nbooktitle = {Proceedings of the AAAI Conference on Artificial Intelligence},<br \/>\r\nvolume = {35},<br \/>\r\nnumber = {17},<br \/>\r\npages = {15053--15057},<br \/>\r\nabstract = {The ability to learn and reason with causal knowledge is a key aspect of intelligent behavior. In contrast to mere statistical association, knowledge of causation enables reasoning about the effects of actions. Causal reasoning is vital for autonomous agents and for a range of applications in science, medicine, business, and government. However, current methods for causal inference are hobbled because they use relatively inexpressive models. Surprisingly, current causal models eschew nearly every major representational innovation common in a range of other fields both inside and outside of computer science, including representation of objects, relationships, time, space, and hierarchy. Even more surprisingly, a range of recent research provides strong evidence that more expressive representations make possible causal inferences that are otherwise impossible and remove key biases that would otherwise afflict more naive inferences. New research on causal inference should target increases in expressiveness to improve accuracy and effectiveness.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('736','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_736\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The ability to learn and reason with causal knowledge is a key aspect of intelligent behavior. In contrast to mere statistical association, knowledge of causation enables reasoning about the effects of actions. Causal reasoning is vital for autonomous agents and for a range of applications in science, medicine, business, and government. However, current methods for causal inference are hobbled because they use relatively inexpressive models. Surprisingly, current causal models eschew nearly every major representational innovation common in a range of other fields both inside and outside of computer science, including representation of objects, relationships, time, space, and hierarchy. Even more surprisingly, a range of recent research provides strong evidence that more expressive representations make possible causal inferences that are otherwise impossible and remove key biases that would otherwise afflict more naive inferences. New research on causal inference should target increases in expressiveness to improve accuracy and effectiveness.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('736','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_736\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/www.aaai.org\/AAAI21Papers\/SMT-427.JensenD.pdf\" title=\"https:\/\/www.aaai.org\/AAAI21Papers\/SMT-427.JensenD.pdf\" target=\"_blank\">https:\/\/www.aaai.org\/AAAI21Papers\/SMT-427.JensenD.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('736','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Akanksha Atrey, Prashant J. Shenoy, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('734','tp_abstract')\" style=\"cursor:pointer;\">Preserving Privacy in Personalized Models for Distributed Mobile Services<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2021<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_734\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('734','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_734\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('734','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_734\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('734','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_734\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/abs-2101-05855,<br \/>\r\ntitle = {Preserving Privacy in Personalized Models for Distributed Mobile Services},<br \/>\r\nauthor = {Akanksha Atrey and Prashant J. Shenoy and David Jensen},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2101.05855},<br \/>\r\nyear  = {2021},<br \/>\r\ndate = {2021-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/2101.05855},<br \/>\r\nabstract = {The ubiquity of mobile devices has led to the proliferation of mobile services that provide personalized and context-aware content to their users. Modern mobile services are distributed between end-devices, such as smartphones, and remote servers that reside in the cloud. Such services thrive on their ability to predict future contexts to pre-fetch content or make context-specific recommendations. An increasingly common method to predict future contexts, such as location, is via machine learning (ML) models. Recent work in context prediction has focused on ML model personalization where a personalized model is learned for each individual user in order to tailor predictions or recommendations to a user's mobile behavior. While the use of personalized models increases efficacy of the mobile service, we argue that it increases privacy risk since a personalized model encodes contextual behavior unique to each user. To demonstrate these privacy risks, we present several attribute inference-based privacy attacks and show that such attacks can leak privacy with up to 78% efficacy for top-3 predictions. We present Pelican, a privacy-preserving personalization system for context-aware mobile services that leverages both device and cloud resources to personalize ML models while minimizing the risk of privacy leakage for users. We evaluate Pelican using real world traces for location-aware mobile services and show that Pelican can substantially reduce privacy leakage by up to 75%.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('734','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_734\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The ubiquity of mobile devices has led to the proliferation of mobile services that provide personalized and context-aware content to their users. Modern mobile services are distributed between end-devices, such as smartphones, and remote servers that reside in the cloud. Such services thrive on their ability to predict future contexts to pre-fetch content or make context-specific recommendations. An increasingly common method to predict future contexts, such as location, is via machine learning (ML) models. Recent work in context prediction has focused on ML model personalization where a personalized model is learned for each individual user in order to tailor predictions or recommendations to a user's mobile behavior. While the use of personalized models increases efficacy of the mobile service, we argue that it increases privacy risk since a personalized model encodes contextual behavior unique to each user. To demonstrate these privacy risks, we present several attribute inference-based privacy attacks and show that such attacks can leak privacy with up to 78% efficacy for top-3 predictions. We present Pelican, a privacy-preserving personalization system for context-aware mobile services that leverages both device and cloud resources to personalize ML models while minimizing the risk of privacy leakage for users. We evaluate Pelican using real world traces for location-aware mobile services and show that Pelican can substantially reduce privacy leakage by up to 75%.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('734','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_734\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2101.05855\" title=\"https:\/\/arxiv.org\/abs\/2101.05855\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2101.05855<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('734','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Sam Witty, David Jensen, Vikash Mansinghka<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('735','tp_abstract')\" style=\"cursor:pointer;\">A Simulation-Based Test of Identifiability for Bayesian Causal Inference<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2021<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_735\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('735','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_735\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('735','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_735\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('735','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_735\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/abs-2102-11761,<br \/>\r\ntitle = {A Simulation-Based Test of Identifiability for Bayesian Causal Inference},<br \/>\r\nauthor = {Sam Witty and David Jensen and Vikash Mansinghka},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2102.11761},<br \/>\r\nyear  = {2021},<br \/>\r\ndate = {2021-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/2102.11761},<br \/>\r\nabstract = {This paper introduces a procedure for testing the identifiability of Bayesian models for causal inference. Although the do-calculus is sound and complete given a causal graph, many practical assumptions cannot be expressed in terms of graph structure alone, such as the assumptions required by instrumental variable designs, regression discontinuity designs, and within-subjects designs. We present simulation-based identifiability (SBI), a fully automated identification test based on a particle optimization scheme with simulated observations. This approach expresses causal assumptions as priors over functions in a structural causal model, including flexible priors using Gaussian processes. We prove that SBI is asymptotically sound and complete, and produces practical finite-sample bounds. We also show empirically that SBI agrees with known results in graph-based identification as well as with widely-held intuitions for designs in which graph-based methods are inconclusive.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('735','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_735\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This paper introduces a procedure for testing the identifiability of Bayesian models for causal inference. Although the do-calculus is sound and complete given a causal graph, many practical assumptions cannot be expressed in terms of graph structure alone, such as the assumptions required by instrumental variable designs, regression discontinuity designs, and within-subjects designs. We present simulation-based identifiability (SBI), a fully automated identification test based on a particle optimization scheme with simulated observations. This approach expresses causal assumptions as priors over functions in a structural causal model, including flexible priors using Gaussian processes. We prove that SBI is asymptotically sound and complete, and produces practical finite-sample bounds. We also show empirically that SBI agrees with known results in graph-based identification as well as with widely-held intuitions for designs in which graph-based methods are inconclusive.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('735','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_735\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2102.11761\" title=\"https:\/\/arxiv.org\/abs\/2102.11761\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2102.11761<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('735','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2020\">2020<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Ifeoluwa Adelani, Ryota Kobayashi, Ingmar Weber, Przemyslaw A. Grabowicz<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('843','tp_abstract')\" style=\"cursor:pointer;\">Estimating community feedback effect on topic choice in social media with predictive modeling<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">EPJ Data Science, <\/span><span class=\"tp_pub_additional_volume\">vol. 9, <\/span><span class=\"tp_pub_additional_number\">no. 1, <\/span><span class=\"tp_pub_additional_pages\">pp. 25, <\/span><span class=\"tp_pub_additional_year\">2020<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 2193-1127<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_843\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('843','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_843\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('843','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_843\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('843','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_843\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Adelani2020,<br \/>\r\ntitle = {Estimating community feedback effect on topic choice in social media with predictive modeling},<br \/>\r\nauthor = {David Ifeoluwa Adelani and Ryota Kobayashi and Ingmar Weber and Przemyslaw A. Grabowicz},<br \/>\r\nurl = {http:\/\/dx.doi.org\/10.1140\/epjds\/s13688-020-00243-w https:\/\/epjdatascience.springeropen.com\/articles\/10.1140\/epjds\/s13688-020-00243-w},<br \/>\r\ndoi = {10.1140\/epjds\/s13688-020-00243-w},<br \/>\r\nissn = {2193-1127},<br \/>\r\nyear  = {2020},<br \/>\r\ndate = {2020-12-01},<br \/>\r\njournal = {EPJ Data Science},<br \/>\r\nvolume = {9},<br \/>\r\nnumber = {1},<br \/>\r\npages = {25},<br \/>\r\npublisher = {The Author(s)},<br \/>\r\nabstract = {Social media users post content on various topics. A defining feature of social media is that other users can provide feedback\u2014called community feedback\u2014to their content in the form of comments, replies, and retweets. We hypothesize that the amount of received feedback influences the choice of topics on which a social media user posts. However, it is challenging to test this hypothesis as user heterogeneity and external confounders complicate measuring the feedback effect. Here, we investigate this hypothesis with a predictive approach based on an interpretable model of an author's decision to continue the topic of their previous post. We explore the confounding factors, including author's topic preferences and unobserved external factors such as news and social events, by optimizing the predictive accuracy. This approach enables us to identify which users are susceptible to community feedback. Overall, we find that 33% and 14% of active users in Reddit and Twitter, respectively, are influenced by community feedback. The model suggests that this feedback alters the probability of topic continuation up to 14%, depending on the user and the amount of feedback.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('843','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_843\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Social media users post content on various topics. A defining feature of social media is that other users can provide feedback\u2014called community feedback\u2014to their content in the form of comments, replies, and retweets. We hypothesize that the amount of received feedback influences the choice of topics on which a social media user posts. However, it is challenging to test this hypothesis as user heterogeneity and external confounders complicate measuring the feedback effect. Here, we investigate this hypothesis with a predictive approach based on an interpretable model of an author's decision to continue the topic of their previous post. We explore the confounding factors, including author's topic preferences and unobserved external factors such as news and social events, by optimizing the predictive accuracy. This approach enables us to identify which users are susceptible to community feedback. Overall, we find that 33% and 14% of active users in Reddit and Twitter, respectively, are influenced by community feedback. The model suggests that this feedback alters the probability of topic continuation up to 14%, depending on the user and the amount of feedback.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('843','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_843\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/dx.doi.org\/10.1140\/epjds\/s13688-020-00243-w https:\/\/epjdatascience.springeropen.com\/articles\/10.1140\/epjds\/s13688-020-00243-w\" title=\"http:\/\/dx.doi.org\/10.1140\/epjds\/s13688-020-00243-w https:\/\/epjdatascience.spring[...]\" target=\"_blank\">http:\/\/dx.doi.org\/10.1140\/epjds\/s13688-020-00243-w https:\/\/epjdatascience.spring[...]<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1140\/epjds\/s13688-020-00243-w\" title=\"Follow DOI:10.1140\/epjds\/s13688-020-00243-w\" target=\"_blank\">doi:10.1140\/epjds\/s13688-020-00243-w<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('843','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Jensen, Javier Burroni, Matthew Rattigan<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('726','tp_abstract')\" style=\"cursor:pointer;\">Object conditioning for causal inference<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Uncertainty in Artificial Intelligence, <\/span><span class=\"tp_pub_additional_pages\">pp. 1072\u20131082, <\/span><span class=\"tp_pub_additional_organization\">PMLR <\/span><span class=\"tp_pub_additional_year\">2020<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_726\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('726','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_726\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('726','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_726\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('726','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_726\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{jensen2020object,<br \/>\r\ntitle = {Object conditioning for causal inference},<br \/>\r\nauthor = {David Jensen and Javier Burroni and Matthew Rattigan},<br \/>\r\nurl = {http:\/\/proceedings.mlr.press\/v115\/jensen20a\/jensen20a.pdf},<br \/>\r\nyear  = {2020},<br \/>\r\ndate = {2020-01-01},<br \/>\r\nbooktitle = {Uncertainty in Artificial Intelligence},<br \/>\r\npages = {1072--1082},<br \/>\r\norganization = {PMLR},<br \/>\r\nabstract = {We describe and analyze a form of conditioning that is widely applied within social science and applied statistics but that is virtually unknown within causal graphical models. This approach, which we term object conditioning, can adjust for the effects of latent confounders and yet avoid the pitfall of conditioning on colliders. We describe object conditioning using plate models and show how its probabilistic implications can be explained using the property of exchangeability. We show that several seemingly obvious interpretations of object conditioning are insufficient to describe its probabilistic implications. Finally, we use object conditioning to describe and unify key aspects of a diverse set of techniques for causal inference, including within-subjects designs, difference-in-differences designs, and interrupted time-series designs.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('726','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_726\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We describe and analyze a form of conditioning that is widely applied within social science and applied statistics but that is virtually unknown within causal graphical models. This approach, which we term object conditioning, can adjust for the effects of latent confounders and yet avoid the pitfall of conditioning on colliders. We describe object conditioning using plate models and show how its probabilistic implications can be explained using the property of exchangeability. We show that several seemingly obvious interpretations of object conditioning are insufficient to describe its probabilistic implications. Finally, we use object conditioning to describe and unify key aspects of a diverse set of techniques for causal inference, including within-subjects designs, difference-in-differences designs, and interrupted time-series designs.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('726','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_726\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/proceedings.mlr.press\/v115\/jensen20a\/jensen20a.pdf\" title=\"http:\/\/proceedings.mlr.press\/v115\/jensen20a\/jensen20a.pdf\" target=\"_blank\">http:\/\/proceedings.mlr.press\/v115\/jensen20a\/jensen20a.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('726','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Akanksha Atrey, Kaleigh Clary, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('730','tp_abstract')\" style=\"cursor:pointer;\">Exploratory Not Explanatory: Counterfactual Analysis of Saliency Maps for Deep Reinforcement Learning<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">International Conference on Learning Representations, <\/span><span class=\"tp_pub_additional_year\">2020<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_730\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('730','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_730\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('730','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_730\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('730','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_730\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{atrey2020exploratory,<br \/>\r\ntitle = {Exploratory Not Explanatory: Counterfactual Analysis of Saliency Maps for Deep Reinforcement Learning},<br \/>\r\nauthor = {Akanksha Atrey and Kaleigh Clary and David Jensen},<br \/>\r\nurl = {https:\/\/openreview.net\/pdf?id=rkl3m1BFDB},<br \/>\r\nyear  = {2020},<br \/>\r\ndate = {2020-01-01},<br \/>\r\nbooktitle = {International Conference on Learning Representations},<br \/>\r\nabstract = {Saliency maps are frequently used to support explanations of the behavior of deep reinforcement learning (RL) agents. However, a review of how saliency maps are used in practice indicates that the derived explanations are often unfalsifiable and can be highly subjective. We introduce an empirical approach grounded in counterfactual reasoning to test the hypotheses generated from saliency maps and assess the degree to which they correspond to the semantics of RL environments. We use Atari games, a common benchmark for deep RL, to evaluate three types of saliency maps. Our results show the extent to which existing claims about Atari games can be evaluated and suggest that saliency maps are best viewed as an exploratory tool rather than an explanatory tool.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('730','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_730\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Saliency maps are frequently used to support explanations of the behavior of deep reinforcement learning (RL) agents. However, a review of how saliency maps are used in practice indicates that the derived explanations are often unfalsifiable and can be highly subjective. We introduce an empirical approach grounded in counterfactual reasoning to test the hypotheses generated from saliency maps and assess the degree to which they correspond to the semantics of RL environments. We use Atari games, a common benchmark for deep RL, to evaluate three types of saliency maps. Our results show the extent to which existing claims about Atari games can be evaluated and suggest that saliency maps are best viewed as an exploratory tool rather than an explanatory tool.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('730','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_730\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/openreview.net\/pdf?id=rkl3m1BFDB\" title=\"https:\/\/openreview.net\/pdf?id=rkl3m1BFDB\" target=\"_blank\">https:\/\/openreview.net\/pdf?id=rkl3m1BFDB<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('730','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Katherine A. Keith, David Jensen, Brendan O'Connor<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('731','tp_abstract')\" style=\"cursor:pointer;\">Text and Causal Inference: A Review of Using Text to Remove Confounding \r\n from Causal Estimates<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 58th Annual Meeting of the Association for Computational \r\n Linguistics, ACL 2020, Online, July 5-10, 2020, <\/span><span class=\"tp_pub_additional_pages\">pp. 5332\u20135344, <\/span><span class=\"tp_pub_additional_publisher\">Association for Computational Linguistics, <\/span><span class=\"tp_pub_additional_year\">2020<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_731\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('731','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_731\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('731','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_731\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('731','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_731\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/acl\/KeithJO20,<br \/>\r\ntitle = {Text and Causal Inference: A Review of Using Text to Remove Confounding <br \/>\r\n from Causal Estimates},<br \/>\r\nauthor = {Katherine A. Keith and David Jensen and Brendan O'Connor},<br \/>\r\nurl = {https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.474},<br \/>\r\nyear  = {2020},<br \/>\r\ndate = {2020-01-01},<br \/>\r\nbooktitle = {Proceedings of the 58th Annual Meeting of the Association for Computational <br \/>\r\n Linguistics, ACL 2020, Online, July 5-10, 2020},<br \/>\r\npages = {5332--5344},<br \/>\r\npublisher = {Association for Computational Linguistics},<br \/>\r\nabstract = {Many applications of computational social science aim to infer causal conclusions from non-experimental data. Such observational data often contains confounders, variables that influence both potential causes and potential effects. Unmeasured or latent confounders can bias causal estimates, and this has motivated interest in measuring potential confounders from observed text. For example, an individual's entire history of social media posts or the content of a news article could provide a rich measurement of multiple confounders. Yet, methods and applications for this problem are scattered across different communities and evaluation practices are inconsistent. This review is the first to gather and categorize these examples and provide a guide to data-processing and evaluation decisions. Despite increased attention on adjusting for confounding using text, there are still many open problems, which we highlight in this paper.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('731','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_731\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Many applications of computational social science aim to infer causal conclusions from non-experimental data. Such observational data often contains confounders, variables that influence both potential causes and potential effects. Unmeasured or latent confounders can bias causal estimates, and this has motivated interest in measuring potential confounders from observed text. For example, an individual's entire history of social media posts or the content of a news article could provide a rich measurement of multiple confounders. Yet, methods and applications for this problem are scattered across different communities and evaluation practices are inconsistent. This review is the first to gather and categorize these examples and provide a guide to data-processing and evaluation decisions. Despite increased attention on adjusting for confounding using text, there are still many open problems, which we highlight in this paper.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('731','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_731\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.474\" title=\"https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.474\" target=\"_blank\">https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.474<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('731','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Sam Witty, Kenta Takatsu, David Jensen, Vikash Mansinghka<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('732','tp_abstract')\" style=\"cursor:pointer;\">Causal Inference using Gaussian Processes with Structured Latent Confounders<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 37th International Conference on Machine Learning, \r\n ICML 2020, 13-18 July 2020, Virtual Event, <\/span><span class=\"tp_pub_additional_pages\">pp. 10313\u201310323, <\/span><span class=\"tp_pub_additional_publisher\">PMLR, <\/span><span class=\"tp_pub_additional_year\">2020<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_732\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('732','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_732\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('732','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_732\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('732','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_732\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/icml\/WittyTJM20,<br \/>\r\ntitle = {Causal Inference using Gaussian Processes with Structured Latent Confounders},<br \/>\r\nauthor = {Sam Witty and Kenta Takatsu and David Jensen and Vikash Mansinghka},<br \/>\r\nurl = {http:\/\/proceedings.mlr.press\/v119\/witty20a.html},<br \/>\r\nyear  = {2020},<br \/>\r\ndate = {2020-01-01},<br \/>\r\nbooktitle = {Proceedings of the 37th International Conference on Machine Learning, <br \/>\r\n ICML 2020, 13-18 July 2020, Virtual Event},<br \/>\r\nvolume = {119},<br \/>\r\npages = {10313--10323},<br \/>\r\npublisher = {PMLR},<br \/>\r\nseries = {Proceedings of Machine Learning Research},<br \/>\r\nabstract = {Latent confounders---unobserved variables that influence both treatment and outcome---can bias estimates of causal effects. In some cases, these confounders are shared across observations, e.g. all students taking a course are influenced by the course's difficulty in addition to any educational interventions they receive individually. This paper shows how to semiparametrically model latent confounders that have this structure and thereby improve estimates of causal effects. The key innovations are a hierarchical Bayesian model, Gaussian processes with structured latent confounders (GP-SLC), and a Monte Carlo inference algorithm for this model based on elliptical slice sampling. GP-SLC provides principled Bayesian uncertainty estimates of individual treatment effect with minimal assumptions about the functional forms relating confounders, covariates, treatment, and outcome. Finally, this paper shows GP-SLC is competitive with or more accurate than widely used causal inference techniques on three benchmark datasets, including the Infant Health and Development Program and a dataset showing the effect of changing temperatures on state-wide energy consumption across New England.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('732','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_732\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Latent confounders---unobserved variables that influence both treatment and outcome---can bias estimates of causal effects. In some cases, these confounders are shared across observations, e.g. all students taking a course are influenced by the course's difficulty in addition to any educational interventions they receive individually. This paper shows how to semiparametrically model latent confounders that have this structure and thereby improve estimates of causal effects. The key innovations are a hierarchical Bayesian model, Gaussian processes with structured latent confounders (GP-SLC), and a Monte Carlo inference algorithm for this model based on elliptical slice sampling. GP-SLC provides principled Bayesian uncertainty estimates of individual treatment effect with minimal assumptions about the functional forms relating confounders, covariates, treatment, and outcome. Finally, this paper shows GP-SLC is competitive with or more accurate than widely used causal inference techniques on three benchmark datasets, including the Infant Health and Development Program and a dataset showing the effect of changing temperatures on state-wide energy consumption across New England.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('732','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_732\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/proceedings.mlr.press\/v119\/witty20a.html\" title=\"http:\/\/proceedings.mlr.press\/v119\/witty20a.html\" target=\"_blank\">http:\/\/proceedings.mlr.press\/v119\/witty20a.html<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('732','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Amanda Gentzel, Justin Clarke, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('733','tp_abstract')\" style=\"cursor:pointer;\">Using Experimental Data to Evaluate Methods for Observational Causal \r\n Inference<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2020<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_733\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('733','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_733\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('733','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_733\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('733','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_733\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/abs-2010-03051,<br \/>\r\ntitle = {Using Experimental Data to Evaluate Methods for Observational Causal <br \/>\r\n Inference},<br \/>\r\nauthor = {Amanda Gentzel and Justin Clarke and David Jensen},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/2010.03051},<br \/>\r\nyear  = {2020},<br \/>\r\ndate = {2020-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/2010.03051},<br \/>\r\nabstract = {Methods that infer causal dependence from observational data are central to many areas of science, including medicine, economics, and the social sciences. A variety of theoretical properties of these methods have been proven, but empirical evaluation remains a challenge, largely due to the lack of observational data sets for which treatment effect is known. We propose and analyze observational sampling from randomized controlled trials (OSRCT), a method for evaluating causal inference methods using data from randomized controlled trials (RCTs). This method can be used to create constructed observational data sets with corresponding unbiased estimates of treatment effect, substantially increasing the number of data sets available for evaluating causal inference methods. We show that, in expectation, OSRCT creates data sets that are equivalent to those produced by randomly sampling from empirical data sets in which all potential outcomes are available. We analyze several properties of OSRCT theoretically and empirically, and we demonstrate its use by comparing the performance of four causal inference methods using data from eleven RCTs.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('733','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_733\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Methods that infer causal dependence from observational data are central to many areas of science, including medicine, economics, and the social sciences. A variety of theoretical properties of these methods have been proven, but empirical evaluation remains a challenge, largely due to the lack of observational data sets for which treatment effect is known. We propose and analyze observational sampling from randomized controlled trials (OSRCT), a method for evaluating causal inference methods using data from randomized controlled trials (RCTs). This method can be used to create constructed observational data sets with corresponding unbiased estimates of treatment effect, substantially increasing the number of data sets available for evaluating causal inference methods. We show that, in expectation, OSRCT creates data sets that are equivalent to those produced by randomly sampling from empirical data sets in which all potential outcomes are available. We analyze several properties of OSRCT theoretically and empirically, and we demonstrate its use by comparing the performance of four causal inference methods using data from eleven RCTs.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('733','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_733\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/2010.03051\" title=\"https:\/\/arxiv.org\/abs\/2010.03051\" target=\"_blank\">https:\/\/arxiv.org\/abs\/2010.03051<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('733','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2019\">2019<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Przemyslaw A. Grabowicz, Nicholas Perello, Kenta Takatsu<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('842','tp_abstract')\" style=\"cursor:pointer;\">Resilience of Supervised Learning Algorithms to Discriminatory Data Perturbations<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_842\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('842','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_842\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('842','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_842\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('842','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_842\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{Grabowicz2019c,<br \/>\r\ntitle = {Resilience of Supervised Learning Algorithms to Discriminatory Data Perturbations},<br \/>\r\nauthor = {Przemyslaw A. Grabowicz and Nicholas Perello and Kenta Takatsu},<br \/>\r\nurl = {http:\/\/arxiv.org\/abs\/1912.08189},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-12-01},<br \/>\r\nabstract = {Discrimination is a focal concern in supervised learning algorithms augmenting human decision-making. These systems are trained using historical data, which may have been tainted by discrimination, and may learn biases against the protected groups. An important question is how to train models without propagating discrimination. In this study, we i) define and model discrimination as perturbations of a data-generating process and show how discrimination can be induced via attributes correlated with the protected attributes; ii) introduce a measure of resilience of a supervised learning algorithm to potentially discriminatory data perturbations, iii) propose a novel supervised learning algorithm that inhibits discrimination, and iv) show that it is more resilient to discriminatory perturbations in synthetic and real-world datasets than state-of-the-art learning algorithms. The proposed method can be used with general supervised learning algorithms and avoids inducement of discrimination, while maximizing model accuracy.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('842','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_842\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Discrimination is a focal concern in supervised learning algorithms augmenting human decision-making. These systems are trained using historical data, which may have been tainted by discrimination, and may learn biases against the protected groups. An important question is how to train models without propagating discrimination. In this study, we i) define and model discrimination as perturbations of a data-generating process and show how discrimination can be induced via attributes correlated with the protected attributes; ii) introduce a measure of resilience of a supervised learning algorithm to potentially discriminatory data perturbations, iii) propose a novel supervised learning algorithm that inhibits discrimination, and iv) show that it is more resilient to discriminatory perturbations in synthetic and real-world datasets than state-of-the-art learning algorithms. The proposed method can be used with general supervised learning algorithms and avoids inducement of discrimination, while maximizing model accuracy.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('842','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_842\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/arxiv.org\/abs\/1912.08189\" title=\"http:\/\/arxiv.org\/abs\/1912.08189\" target=\"_blank\">http:\/\/arxiv.org\/abs\/1912.08189<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('842','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Jensen, others<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('778','tp_abstract')\" style=\"cursor:pointer;\">Comment: Strengthening empirical evaluation of causal inference methods<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Statistical Science, <\/span><span class=\"tp_pub_additional_volume\">vol. 34, <\/span><span class=\"tp_pub_additional_number\">no. 1, <\/span><span class=\"tp_pub_additional_pages\">pp. 77\u201381, <\/span><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_778\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('778','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_778\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('778','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_778\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('778','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_778\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{jensen2019comment,<br \/>\r\ntitle = {Comment: Strengthening empirical evaluation of causal inference methods},<br \/>\r\nauthor = {David Jensen and others},<br \/>\r\nurl = {https:\/\/projecteuclid.org\/journals\/statistical-science\/volume-34\/issue-1\/Comment-Strengthening-Empirical-Evaluation-of-Causal-Inference-Methods\/10.1214\/18-STS690.short},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-01-01},<br \/>\r\njournal = {Statistical Science},<br \/>\r\nvolume = {34},<br \/>\r\nnumber = {1},<br \/>\r\npages = {77--81},<br \/>\r\npublisher = {Institute of Mathematical Statistics},<br \/>\r\nabstract = {This is a contribution to the discussion of the paper by Dorie et al. (Statist. Sci. 34 (2019) 43\u201368), which reports the lessons learned from 2016 Atlantic Causal Inference Conference Competition. My comments strongly support the authors\u2019 focus on empirical evaluation, using examples and experience from machine learning research, particularly focusing on the problem of algorithmic complexity. I argue that even broader and deeper empirical evaluation should be undertaken by the researchers who study causal inference. Finally, I highlight a few key conclusions that suggest where future research should focus.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('778','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_778\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This is a contribution to the discussion of the paper by Dorie et al. (Statist. Sci. 34 (2019) 43\u201368), which reports the lessons learned from 2016 Atlantic Causal Inference Conference Competition. My comments strongly support the authors\u2019 focus on empirical evaluation, using examples and experience from machine learning research, particularly focusing on the problem of algorithmic complexity. I argue that even broader and deeper empirical evaluation should be undertaken by the researchers who study causal inference. Finally, I highlight a few key conclusions that suggest where future research should focus.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('778','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_778\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/projecteuclid.org\/journals\/statistical-science\/volume-34\/issue-1\/Comment-Strengthening-Empirical-Evaluation-of-Causal-Inference-Methods\/10.1214\/18-STS690.short\" title=\"https:\/\/projecteuclid.org\/journals\/statistical-science\/volume-34\/issue-1\/Comment[...]\" target=\"_blank\">https:\/\/projecteuclid.org\/journals\/statistical-science\/volume-34\/issue-1\/Comment[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('778','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Emma Tosch, Eytan Bakshy, Emery D Berger, David Jensen, J Eliot B Moss<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('727','tp_abstract')\" style=\"cursor:pointer;\">PlanAlyzer: assessing threats to the validity of online experiments<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Proceedings of the ACM on Programming Languages, <\/span><span class=\"tp_pub_additional_volume\">vol. 3, <\/span><span class=\"tp_pub_additional_number\">no. OOPSLA, <\/span><span class=\"tp_pub_additional_pages\">pp. 1\u201330, <\/span><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_727\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('727','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_727\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('727','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_727\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('727','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_727\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{tosch2019planalyzer,<br \/>\r\ntitle = {PlanAlyzer: assessing threats to the validity of online experiments},<br \/>\r\nauthor = {Emma Tosch and Eytan Bakshy and Emery D Berger and David Jensen and J Eliot B Moss},<br \/>\r\nurl = {https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3360608},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-01-01},<br \/>\r\njournal = {Proceedings of the ACM on Programming Languages},<br \/>\r\nvolume = {3},<br \/>\r\nnumber = {OOPSLA},<br \/>\r\npages = {1--30},<br \/>\r\npublisher = {ACM New York, NY, USA},<br \/>\r\nabstract = {Online experiments are ubiquitous. As the scale of experiments has grown, so has the complexity of their design and implementation. In response, firms have developed software frameworks for designing and deploying online experiments. Ensuring that experiments in these frameworks are correctly designed and that their results are trustworthy---referred to as *internal validity*---can be difficult. Currently, verifying internal validity requires manual inspection by someone with substantial expertise in experimental design. We present the first approach for statically checking the internal validity of online experiments. Our checks are based on well-known problems that arise in experimental design and causal inference. Our analyses target PlanOut, a widely deployed, open-source experimentation framework that uses a domain-specific language to specify and run complex experiments. We have built a tool, PlanAlyzer, that checks PlanOut programs for a variety of threats to internal validity, including failures of randomization, treatment assignment, and causal sufficiency. PlanAlyzer uses its analyses to automatically generate *contrasts*, a key type of information required to perform valid statistical analyses over experimental results. We demonstrate PlanAlyzer's utility on a corpus of PlanOut scripts deployed in production at Facebook, and we evaluate its ability to identify threats to validity on a mutated subset of this corpus. PlanAlyzer has both precision and recall of 92% on the mutated corpus, and 82% of the contrasts it automatically generates match hand-specified data.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('727','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_727\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Online experiments are ubiquitous. As the scale of experiments has grown, so has the complexity of their design and implementation. In response, firms have developed software frameworks for designing and deploying online experiments. Ensuring that experiments in these frameworks are correctly designed and that their results are trustworthy---referred to as *internal validity*---can be difficult. Currently, verifying internal validity requires manual inspection by someone with substantial expertise in experimental design. We present the first approach for statically checking the internal validity of online experiments. Our checks are based on well-known problems that arise in experimental design and causal inference. Our analyses target PlanOut, a widely deployed, open-source experimentation framework that uses a domain-specific language to specify and run complex experiments. We have built a tool, PlanAlyzer, that checks PlanOut programs for a variety of threats to internal validity, including failures of randomization, treatment assignment, and causal sufficiency. PlanAlyzer uses its analyses to automatically generate *contrasts*, a key type of information required to perform valid statistical analyses over experimental results. We demonstrate PlanAlyzer's utility on a corpus of PlanOut scripts deployed in production at Facebook, and we evaluate its ability to identify threats to validity on a mutated subset of this corpus. PlanAlyzer has both precision and recall of 92% on the mutated corpus, and 82% of the contrasts it automatically generates match hand-specified data.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('727','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_727\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3360608\" title=\"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3360608\" target=\"_blank\">https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3360608<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('727','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Huseyin Oktay, Akanksha Atrey, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('724','tp_abstract')\" style=\"cursor:pointer;\">Identifying when effect restoration will improve estimates of causal effect<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2019 SIAM International Conference on Data Mining, <\/span><span class=\"tp_pub_additional_pages\">pp. 190\u2013198, <\/span><span class=\"tp_pub_additional_organization\">Society for Industrial and Applied Mathematics <\/span><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_724\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('724','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_724\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('724','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_724\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('724','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_724\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{oktay2019identifying,<br \/>\r\ntitle = {Identifying when effect restoration will improve estimates of causal effect},<br \/>\r\nauthor = {Huseyin Oktay and Akanksha Atrey and David Jensen},<br \/>\r\nurl = {https:\/\/doi.org\/10.1137\/1.9781611975673.22},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-01-01},<br \/>\r\nbooktitle = {Proceedings of the 2019 SIAM International Conference on Data Mining},<br \/>\r\npages = {190--198},<br \/>\r\norganization = {Society for Industrial and Applied Mathematics},<br \/>\r\nabstract = {Several methods have been developed that combine multiple models learned on different data sets and then use that combination to reach conclusions that would not have been possible with any one of the models alone. We examine one such method\u2014effect restoration\u2014which was originally developed to mitigate the effects of poorly measured confounding variables in a causal model. We show how effect restoration can be used to combine results from different machine learning models and how the combined model can be used to estimate causal effects that are not identifiable from either of the original studies alone. We characterize the performance of effect restoration by using both theoretical analysis and simulation studies. Specifically, we show how conditional independence tests and common assumptions can help distinguish when effect restoration should and should not be applied, and we use empirical analysis to show the limited range of conditions under which effect restoration should be applied in practical situations.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('724','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_724\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Several methods have been developed that combine multiple models learned on different data sets and then use that combination to reach conclusions that would not have been possible with any one of the models alone. We examine one such method\u2014effect restoration\u2014which was originally developed to mitigate the effects of poorly measured confounding variables in a causal model. We show how effect restoration can be used to combine results from different machine learning models and how the combined model can be used to estimate causal effects that are not identifiable from either of the original studies alone. We characterize the performance of effect restoration by using both theoretical analysis and simulation studies. Specifically, we show how conditional independence tests and common assumptions can help distinguish when effect restoration should and should not be applied, and we use empirical analysis to show the limited range of conditions under which effect restoration should be applied in practical situations.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('724','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_724\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.1137\/1.9781611975673.22\" title=\"https:\/\/doi.org\/10.1137\/1.9781611975673.22\" target=\"_blank\">https:\/\/doi.org\/10.1137\/1.9781611975673.22<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('724','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Amanda Gentzel, Dan Garant, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('728','tp_abstract')\" style=\"cursor:pointer;\">The Case for Evaluating Causal Models Using Interventional Measures and Empirical Data<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Advances in Neural Information Processing Systems, <\/span><span class=\"tp_pub_additional_publisher\">Curran Associates, Inc., <\/span><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_728\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('728','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_728\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('728','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_728\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('728','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_728\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{gentzel2019case,<br \/>\r\ntitle = {The Case for Evaluating Causal Models Using Interventional Measures and Empirical Data},<br \/>\r\nauthor = {Amanda Gentzel and Dan Garant and David Jensen},<br \/>\r\nurl = {https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/a87c11b9100c608b7f8e98cfa316ff7b-Paper.pdf},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-01-01},<br \/>\r\nbooktitle = {Advances in Neural Information Processing Systems},<br \/>\r\nvolume = {32},<br \/>\r\npublisher = {Curran Associates, Inc.},<br \/>\r\nabstract = {Causal inference is central to many areas of artificial intelligence, including complex reasoning, planning, knowledge-base construction, robotics, explanation, and fairness. An active community of researchers develops and enhances algorithms that learn causal models from data, and this work has produced a series of impressive technical advances. However, evaluation techniques for causal modeling algorithms have remained somewhat primitive, limiting what we can learn from experimental studies of algorithm performance, constraining the types of algorithms and model representations that researchers consider, and creating a gap between theory and practice. We argue for more frequent use of evaluation techniques that examine interventional measures rather than structural or observational measures, and that evaluate those measures on empirical data rather than synthetic data. We survey the current practice in evaluation and show that the techniques we recommend are rarely used in practice. We show that such techniques are feasible and that data sets are available to conduct such evaluations. We also show that these techniques produce substantially different results than using structural measures and synthetic data.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('728','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_728\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Causal inference is central to many areas of artificial intelligence, including complex reasoning, planning, knowledge-base construction, robotics, explanation, and fairness. An active community of researchers develops and enhances algorithms that learn causal models from data, and this work has produced a series of impressive technical advances. However, evaluation techniques for causal modeling algorithms have remained somewhat primitive, limiting what we can learn from experimental studies of algorithm performance, constraining the types of algorithms and model representations that researchers consider, and creating a gap between theory and practice. We argue for more frequent use of evaluation techniques that examine interventional measures rather than structural or observational measures, and that evaluate those measures on empirical data rather than synthetic data. We survey the current practice in evaluation and show that the techniques we recommend are rarely used in practice. We show that such techniques are feasible and that data sets are available to conduct such evaluations. We also show that these techniques produce substantially different results than using structural measures and synthetic data.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('728','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_728\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/a87c11b9100c608b7f8e98cfa316ff7b-Paper.pdf\" title=\"https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/a87c11b9100c608b7f8e98cfa316ff7b-[...]\" target=\"_blank\">https:\/\/proceedings.neurips.cc\/paper\/2019\/file\/a87c11b9100c608b7f8e98cfa316ff7b-[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('728','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Emma Tosch, Kaleigh Clary, John Foley, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('725','tp_abstract')\" style=\"cursor:pointer;\">Toybox: A Suite of Environments for Experimental Evaluation of Deep \r\n Reinforcement Learning<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_725\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('725','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_725\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('725','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_725\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('725','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_725\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/abs-1905-02825,<br \/>\r\ntitle = {Toybox: A Suite of Environments for Experimental Evaluation of Deep <br \/>\r\n Reinforcement Learning},<br \/>\r\nauthor = {Emma Tosch and Kaleigh Clary and John Foley and David Jensen},<br \/>\r\nurl = {http:\/\/arxiv.org\/abs\/1905.02825},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/1905.02825},<br \/>\r\nabstract = {Evaluation of deep reinforcement learning (RL) is inherently challenging. In particular, learned policies are largely opaque, and hypotheses about the behavior of deep RL agents are difficult to test in black-box environments. Considerable effort has gone into addressing opacity, but almost no effort has been devoted to producing high quality environments for experimental evaluation of agent behavior. We present TOYBOX, a new high-performance, open-source* subset of Atari environments re-designed for the experimental evaluation of deep RL. We show that TOYBOX enables a wide range of experiments and analyses that are impossible in other environments.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('725','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_725\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Evaluation of deep reinforcement learning (RL) is inherently challenging. In particular, learned policies are largely opaque, and hypotheses about the behavior of deep RL agents are difficult to test in black-box environments. Considerable effort has gone into addressing opacity, but almost no effort has been devoted to producing high quality environments for experimental evaluation of agent behavior. We present TOYBOX, a new high-performance, open-source* subset of Atari environments re-designed for the experimental evaluation of deep RL. We show that TOYBOX enables a wide range of experiments and analyses that are impossible in other environments.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('725','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_725\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/arxiv.org\/abs\/1905.02825\" title=\"http:\/\/arxiv.org\/abs\/1905.02825\" target=\"_blank\">http:\/\/arxiv.org\/abs\/1905.02825<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('725','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Sam Witty, Alexander Lew, David Jensen, Vikash Mansinghka<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('729','tp_abstract')\" style=\"cursor:pointer;\">Bayesian causal inference via probabilistic program synthesis<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2019<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_729\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('729','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_729\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('729','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_729\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('729','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_729\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{witty2019bayesian,<br \/>\r\ntitle = {Bayesian causal inference via probabilistic program synthesis},<br \/>\r\nauthor = {Sam Witty and Alexander Lew and David Jensen and Vikash Mansinghka},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/1910.14124.pdf},<br \/>\r\nyear  = {2019},<br \/>\r\ndate = {2019-01-01},<br \/>\r\njournal = {arXiv preprint arXiv:1910.14124},<br \/>\r\nabstract = {Causal inference can be formalized as Bayesian inference that combines a prior distribution over causal models and likelihoods that account for both observations and interventions. We show that it is possible to implement this approach using a sufficiently expressive probabilistic programming language. Priors are represented using probabilistic programs that generate source code in a domain specific language. Interventions are represented using probabilistic programs that edit this source code to modify the original generative process. This approach makes it straightforward to incorporate data from atomic interventions, as well as shift interventions, variance-scaling interventions, and other interventions that modify causal structure. This approach also enables the use of general-purpose inference machinery for probabilistic programs to infer probable causal structures and parameters from data. This abstract describes a prototype of this approach in the Gen probabilistic programming language.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('729','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_729\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Causal inference can be formalized as Bayesian inference that combines a prior distribution over causal models and likelihoods that account for both observations and interventions. We show that it is possible to implement this approach using a sufficiently expressive probabilistic programming language. Priors are represented using probabilistic programs that generate source code in a domain specific language. Interventions are represented using probabilistic programs that edit this source code to modify the original generative process. This approach makes it straightforward to incorporate data from atomic interventions, as well as shift interventions, variance-scaling interventions, and other interventions that modify causal structure. This approach also enables the use of general-purpose inference machinery for probabilistic programs to infer probable causal structures and parameters from data. This abstract describes a prototype of this approach in the Gen probabilistic programming language.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('729','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_729\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/1910.14124.pdf\" title=\"https:\/\/arxiv.org\/pdf\/1910.14124.pdf\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/1910.14124.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('729','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2018\">2018<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">John Foley, Emma Tosch, Kaleigh Clary, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('776','tp_abstract')\" style=\"cursor:pointer;\">Toybox: Better Atari Environments for Testing Reinforcement Learning Agents<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">NeurIPS 2018 Workshop on Systems for ML, <\/span><span class=\"tp_pub_additional_year\">2018<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_776\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('776','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_776\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('776','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_776\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('776','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_776\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{foley2018toybox,<br \/>\r\ntitle = {Toybox: Better Atari Environments for Testing Reinforcement Learning Agents},<br \/>\r\nauthor = {John Foley and Emma Tosch and Kaleigh Clary and David Jensen},<br \/>\r\nurl = {http:\/\/learningsys.org\/nips18\/assets\/papers\/83CameraReadySubmissionNIPS_Systems_for_ML_Workshop_2019___ToyBox%20(11).pdf},<br \/>\r\nyear  = {2018},<br \/>\r\ndate = {2018-01-01},<br \/>\r\nbooktitle = {NeurIPS 2018 Workshop on Systems for ML},<br \/>\r\nabstract = {It is a widely accepted principle that software without tests has bugs. Testing reinforcement learning agents is especially difficult because of the stochastic nature of both agents and environments, the complexity of state-of-the-art models,and the sequential nature of their predictions. Recently, the Arcade Learning Environment (ALE) has become one of the most widely used benchmark suites for deep learning research, and state-of-the-art Reinforcement Learning (RL) agents have been shown to routinely equal or exceed human performance on many ALE tasks. Since ALE is based on emulation of original Atari games, the environment does not provide semantically meaningful representations of internal game state. This means that ALE has limited utility as an environment for supporting testing or model introspection. We propose TOYBOX, a collection of reimplementations of these games that solves this critical problem and enables robust testing of RL agents.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('776','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_776\" style=\"display:none;\"><div class=\"tp_abstract_entry\">It is a widely accepted principle that software without tests has bugs. Testing reinforcement learning agents is especially difficult because of the stochastic nature of both agents and environments, the complexity of state-of-the-art models,and the sequential nature of their predictions. Recently, the Arcade Learning Environment (ALE) has become one of the most widely used benchmark suites for deep learning research, and state-of-the-art Reinforcement Learning (RL) agents have been shown to routinely equal or exceed human performance on many ALE tasks. Since ALE is based on emulation of original Atari games, the environment does not provide semantically meaningful representations of internal game state. This means that ALE has limited utility as an environment for supporting testing or model introspection. We propose TOYBOX, a collection of reimplementations of these games that solves this critical problem and enables robust testing of RL agents.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('776','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_776\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/learningsys.org\/nips18\/assets\/papers\/83CameraReadySubmissionNIPS_Systems_for_ML_Workshop_2019___ToyBox%20(11).pdf\" title=\"http:\/\/learningsys.org\/nips18\/assets\/papers\/83CameraReadySubmissionNIPS_Systems_[...]\" target=\"_blank\">http:\/\/learningsys.org\/nips18\/assets\/papers\/83CameraReadySubmissionNIPS_Systems_[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('776','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Sam Witty, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('777','tp_abstract')\" style=\"cursor:pointer;\">Causal Graphs vs. Causal Programs: The Case of Conditional Branching<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">First Conference on Probabilistic Programming (ProbProg), <\/span><span class=\"tp_pub_additional_year\">2018<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_777\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('777','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_777\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('777','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_777\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('777','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_777\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{witty2018causal,<br \/>\r\ntitle = {Causal Graphs vs. Causal Programs: The Case of Conditional Branching},<br \/>\r\nauthor = {Sam Witty and David Jensen},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/2007.07127.pdf},<br \/>\r\nyear  = {2018},<br \/>\r\ndate = {2018-01-01},<br \/>\r\nbooktitle = {First Conference on Probabilistic Programming (ProbProg)},<br \/>\r\nabstract = {We evaluate the performance of graph-based causal discovery algorithms when the generative process is a probabilistic program with conditional branching. Using synthetic experiments, we demonstrate empirically that graph-based causal discovery algorithms are able to learn accurate associational distributions for probabilistic programs with contextsensitive structure, but that those graphs fail to accurately model the effects of interventions on the programs},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('777','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_777\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We evaluate the performance of graph-based causal discovery algorithms when the generative process is a probabilistic program with conditional branching. Using synthetic experiments, we demonstrate empirically that graph-based causal discovery algorithms are able to learn accurate associational distributions for probabilistic programs with contextsensitive structure, but that those graphs fail to accurately model the effects of interventions on the programs<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('777','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_777\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/2007.07127.pdf\" title=\"https:\/\/arxiv.org\/pdf\/2007.07127.pdf\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/2007.07127.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('777','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Sam Witty, Jun Ki Lee, Emma Tosch, Akanksha Atrey, Michael Littman, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('720','tp_abstract')\" style=\"cursor:pointer;\">Measuring and characterizing generalization in deep reinforcement learning<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2018<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_720\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('720','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_720\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('720','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_720\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('720','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_720\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{witty2018measuring,<br \/>\r\ntitle = {Measuring and characterizing generalization in deep reinforcement learning},<br \/>\r\nauthor = {Sam Witty and Jun Ki Lee and Emma Tosch and Akanksha Atrey and Michael Littman and David Jensen},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/1812.02868.pdf},<br \/>\r\nyear  = {2018},<br \/>\r\ndate = {2018-01-01},<br \/>\r\njournal = {arXiv preprint arXiv:1812.02868},<br \/>\r\nabstract = {Deep reinforcement-learning methods have achieved remarkable performance on challenging control tasks. Observations of the resulting behavior give the impression that the agent has constructed a generalized representation that supports insightful action decisions. We re-examine what is meant by generalization in RL, and propose several definitions based on an agent's performance in on-policy, off-policy, and unreachable states. We propose a set of practical methods for evaluating agents with these definitions of generalization. We demonstrate these techniques on a common benchmark task for deep RL, and we show that the learned networks make poor decisions for states that differ only slightly from on-policy states, even though those states are not selected adversarially. Taken together, these results call into question the extent to which deep Q-networks learn generalized representations, and suggest that more experimentation and analysis is necessary before claims of representation learning can be supported.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('720','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_720\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Deep reinforcement-learning methods have achieved remarkable performance on challenging control tasks. Observations of the resulting behavior give the impression that the agent has constructed a generalized representation that supports insightful action decisions. We re-examine what is meant by generalization in RL, and propose several definitions based on an agent's performance in on-policy, off-policy, and unreachable states. We propose a set of practical methods for evaluating agents with these definitions of generalization. We demonstrate these techniques on a common benchmark task for deep RL, and we show that the learned networks make poor decisions for states that differ only slightly from on-policy states, even though those states are not selected adversarially. Taken together, these results call into question the extent to which deep Q-networks learn generalized representations, and suggest that more experimentation and analysis is necessary before claims of representation learning can be supported.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('720','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_720\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/1812.02868.pdf\" title=\"https:\/\/arxiv.org\/pdf\/1812.02868.pdf\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/1812.02868.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('720','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Kaleigh Clary, Emma Tosch, John Foley, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('722','tp_abstract')\" style=\"cursor:pointer;\">Let's Play Again: Variability of Deep Reinforcement Learning Agents in Atari Environments<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2018<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_722\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('722','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_722\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('722','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_722\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('722','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_722\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{clary2018variability,<br \/>\r\ntitle = {Let's Play Again: Variability of Deep Reinforcement Learning Agents in Atari Environments},<br \/>\r\nauthor = {Kaleigh Clary and Emma Tosch and John Foley and David Jensen},<br \/>\r\nurl = {https:\/\/arxiv.org\/pdf\/1904.06312},<br \/>\r\nyear  = {2018},<br \/>\r\ndate = {2018-01-01},<br \/>\r\nbooktitle = {Critiquing and Correcting Trends in Machine Learning Workshop at Neural Information Processing Systems},<br \/>\r\nabstract = {Reproducibility in reinforcement learning is challenging: uncontrolled stochasticity from many sources, such as the learning algorithm, the learned policy, and the environment itself have led researchers to report the performance of learned agents using aggregate metrics of performance over multiple random seeds for a single environment. Unfortunately, there are still pernicious sources of variability in reinforcement learning agents that make reporting common summary statistics an unsound metric for performance. Our experiments demonstrate the variability of common agents used in the popular OpenAI Baselines repository. We make the case for reporting post-training agent performance as a distribution, rather than a point estimate.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('722','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_722\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Reproducibility in reinforcement learning is challenging: uncontrolled stochasticity from many sources, such as the learning algorithm, the learned policy, and the environment itself have led researchers to report the performance of learned agents using aggregate metrics of performance over multiple random seeds for a single environment. Unfortunately, there are still pernicious sources of variability in reinforcement learning agents that make reporting common summary statistics an unsound metric for performance. Our experiments demonstrate the variability of common agents used in the popular OpenAI Baselines repository. We make the case for reporting post-training agent performance as a distribution, rather than a point estimate.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('722','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_722\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/pdf\/1904.06312\" title=\"https:\/\/arxiv.org\/pdf\/1904.06312\" target=\"_blank\">https:\/\/arxiv.org\/pdf\/1904.06312<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('722','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2017\">2017<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Kaleigh Clary, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('774','tp_abstract')\" style=\"cursor:pointer;\">A\/B Testing in Networks with Adversarial Members<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2017<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_774\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('774','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_774\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('774','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_774\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('774','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_774\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{clary2017b,<br \/>\r\ntitle = {A\/B Testing in Networks with Adversarial Members},<br \/>\r\nauthor = {Kaleigh Clary and David Jensen},<br \/>\r\nurl = {http:\/\/www.mlgworkshop.org\/2017\/paper\/MLG2017_paper_27.pdf},<br \/>\r\nyear  = {2017},<br \/>\r\ndate = {2017-01-01},<br \/>\r\nabstract = {Many researchers attempt to study the effects of interventions in network systems. To simplify experimental design and analysis in these environments, simple assumptions are made about the behavior of its members. However, nodes may not respond to treatment, or may respond maliciously. These adversarial nodes influence treatment topology by preventing or altering the expected network effect, but may not be known or detectable. We characterize the influence of adversarial nodes and the bias these nodes introduce in average treatment effect estimates. In particular, we derive expressions for the bias induced in average treatment effect using the linear estimator from Gui et al (2015). In addition to theoretical bounds, we empirically demonstrate estimation bias through experiments on synthetically generated networks. We consider both the case in which adversarial nodes are dispersed randomly through the network and the case where adversarial node placement is targeted to the highest degree nodes. Our work demonstrates that peer influence makes causal estimates on networks susceptible to the actions of adversaries, and specific network structures are particularly vulnerable to to adversarial responses.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('774','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_774\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Many researchers attempt to study the effects of interventions in network systems. To simplify experimental design and analysis in these environments, simple assumptions are made about the behavior of its members. However, nodes may not respond to treatment, or may respond maliciously. These adversarial nodes influence treatment topology by preventing or altering the expected network effect, but may not be known or detectable. We characterize the influence of adversarial nodes and the bias these nodes introduce in average treatment effect estimates. In particular, we derive expressions for the bias induced in average treatment effect using the linear estimator from Gui et al (2015). In addition to theoretical bounds, we empirically demonstrate estimation bias through experiments on synthetically generated networks. We consider both the case in which adversarial nodes are dispersed randomly through the network and the case where adversarial node placement is targeted to the highest degree nodes. Our work demonstrates that peer influence makes causal estimates on networks susceptible to the actions of adversaries, and specific network structures are particularly vulnerable to to adversarial responses.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('774','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_774\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/www.mlgworkshop.org\/2017\/paper\/MLG2017_paper_27.pdf\" title=\"http:\/\/www.mlgworkshop.org\/2017\/paper\/MLG2017_paper_27.pdf\" target=\"_blank\">http:\/\/www.mlgworkshop.org\/2017\/paper\/MLG2017_paper_27.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('774','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Javier Burroni, Arjun Guha, David Jensen<\/p><p class=\"tp_pub_title\">INTERACTIVE WRITING AND DEBUGGING OF BAYESIAN PROBABILISTIC PROGRAMS <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2017<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_775\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('775','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_775\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('775','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_775\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{burroni2017interactive,<br \/>\r\ntitle = {INTERACTIVE WRITING AND DEBUGGING OF BAYESIAN PROBABILISTIC PROGRAMS},<br \/>\r\nauthor = {Javier Burroni and Arjun Guha and David Jensen},<br \/>\r\nurl = {https:\/\/pps2018.luddy.indiana.edu\/files\/2017\/12\/interactive_debugger.pdf},<br \/>\r\nyear  = {2017},<br \/>\r\ndate = {2017-01-01},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('775','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_775\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/pps2018.luddy.indiana.edu\/files\/2017\/12\/interactive_debugger.pdf\" title=\"https:\/\/pps2018.luddy.indiana.edu\/files\/2017\/12\/interactive_debugger.pdf\" target=\"_blank\">https:\/\/pps2018.luddy.indiana.edu\/files\/2017\/12\/interactive_debugger.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('775','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Katerina Marazopoulou, David Arbour, David Jensen<\/p><p class=\"tp_pub_title\">On causal analysis for heterogeneous networks <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">The 2017 ACM SIGKDD Workshop on Causal Discovery, <\/span><span class=\"tp_pub_additional_year\">2017<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_resource_link\"><a id=\"tp_links_sh_779\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('779','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_779\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('779','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_779\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{marazopoulou2017causal,<br \/>\r\ntitle = {On causal analysis for heterogeneous networks},<br \/>\r\nauthor = {Katerina Marazopoulou and David Arbour and David Jensen},<br \/>\r\nurl = {http:\/\/nugget.unisa.edu.au\/CD2017\/slides\/KaterinaMarazopoulou.pdf},<br \/>\r\nyear  = {2017},<br \/>\r\ndate = {2017-01-01},<br \/>\r\nbooktitle = {The 2017 ACM SIGKDD Workshop on Causal Discovery},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('779','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_779\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/nugget.unisa.edu.au\/CD2017\/slides\/KaterinaMarazopoulou.pdf\" title=\"http:\/\/nugget.unisa.edu.au\/CD2017\/slides\/KaterinaMarazopoulou.pdf\" target=\"_blank\">http:\/\/nugget.unisa.edu.au\/CD2017\/slides\/KaterinaMarazopoulou.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('779','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Kaleigh Clary, Andrew McGregor, David Jensen<\/p><p class=\"tp_pub_title\">A\/B Testing in Networks with Adversarial Nodes <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">KDD Workshop on Mining and Learning with Graphs, <\/span><span class=\"tp_pub_additional_year\">2017<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_721\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('721','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_721\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{clary2017adversaries,<br \/>\r\ntitle = {A\/B Testing in Networks with Adversarial Nodes},<br \/>\r\nauthor = {Kaleigh Clary and Andrew McGregor and David Jensen},<br \/>\r\nyear  = {2017},<br \/>\r\ndate = {2017-01-01},<br \/>\r\nbooktitle = {KDD Workshop on Mining and Learning with Graphs},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('721','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2016\">2016<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Arbour, Dan Garant, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('770','tp_abstract')\" style=\"cursor:pointer;\">Inferring Network Effects from Observational Data<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 22nd ACM SIGKDD International Conference on \r\n Knowledge Discovery and Data Mining, San Francisco, CA, USA, August \r\n 13-17, 2016, <\/span><span class=\"tp_pub_additional_pages\">pp. 715\u2013724, <\/span><span class=\"tp_pub_additional_publisher\">ACM, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_770\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('770','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_770\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('770','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_770\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('770','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_770\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/kdd\/ArbourGJ16,<br \/>\r\ntitle = {Inferring Network Effects from Observational Data},<br \/>\r\nauthor = {David Arbour and Dan Garant and David Jensen},<br \/>\r\nurl = {https:\/\/doi.org\/10.1145\/2939672.2939791},<br \/>\r\ndoi = {10.1145\/2939672.2939791},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\nbooktitle = {Proceedings of the 22nd ACM SIGKDD International Conference on <br \/>\r\n Knowledge Discovery and Data Mining, San Francisco, CA, USA, August <br \/>\r\n 13-17, 2016},<br \/>\r\npages = {715--724},<br \/>\r\npublisher = {ACM},<br \/>\r\nabstract = {We present Relational Covariate Adjustment (RCA), a general method for estimating causal effects in relational data. Relational Covariate Adjustment is implemented through two high-level operations: identification of an adjustment set and relational regression adjustment. The former is achieved through an extension of Pearl\u2019s back-door criterion to relational domains. We demonstrate how this extended definition can be used to estimate causal effects in the presence of network interference and confounding. RCA is agnostic to functional form, and it can easily model both discrete and continuous treatments as well as estimate the effects of a wider array of network interventions than existing experimental approaches. We show that RCA can yield robust estimates of causal effects using common regression models without extensive parameter tuning. Through a series of simulation experiments on a variety of synthetic and real- world network structures, we show that causal effects estimated on observational data with RCA are nearly as accurate as those estimated from well-designed network experiments.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('770','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_770\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We present Relational Covariate Adjustment (RCA), a general method for estimating causal effects in relational data. Relational Covariate Adjustment is implemented through two high-level operations: identification of an adjustment set and relational regression adjustment. The former is achieved through an extension of Pearl\u2019s back-door criterion to relational domains. We demonstrate how this extended definition can be used to estimate causal effects in the presence of network interference and confounding. RCA is agnostic to functional form, and it can easily model both discrete and continuous treatments as well as estimate the effects of a wider array of network interventions than existing experimental approaches. We show that RCA can yield robust estimates of causal effects using common regression models without extensive parameter tuning. Through a series of simulation experiments on a variety of synthetic and real- world network structures, we show that causal effects estimated on observational data with RCA are nearly as accurate as those estimated from well-designed network experiments.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('770','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_770\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.1145\/2939672.2939791\" title=\"https:\/\/doi.org\/10.1145\/2939672.2939791\" target=\"_blank\">https:\/\/doi.org\/10.1145\/2939672.2939791<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1145\/2939672.2939791\" title=\"Follow DOI:10.1145\/2939672.2939791\" target=\"_blank\">doi:10.1145\/2939672.2939791<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('770','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Arbour, Katerina Marazopoulou, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('771','tp_abstract')\" style=\"cursor:pointer;\">Inferring Causal Direction from Relational Data<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the Thirty-Second Conference on Uncertainty in Artificial \r\n Intelligence, UAI 2016, June 25-29, 2016, New York City, NY, USA, <\/span><span class=\"tp_pub_additional_publisher\">AUAI Press, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_771\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('771','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_771\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('771','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_771\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('771','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_771\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/uai\/ArbourMJ16,<br \/>\r\ntitle = {Inferring Causal Direction from Relational Data},<br \/>\r\nauthor = {David Arbour and Katerina Marazopoulou and David Jensen},<br \/>\r\nurl = {http:\/\/auai.org\/uai2016\/proceedings\/papers\/217.pdf},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\nbooktitle = {Proceedings of the Thirty-Second Conference on Uncertainty in Artificial <br \/>\r\n Intelligence, UAI 2016, June 25-29, 2016, New York City, NY, USA},<br \/>\r\npublisher = {AUAI Press},<br \/>\r\nabstract = {Inferring the direction of causal dependence from observational data is a fundamental problem in many scientific fields. Significant progress has been made in inferring causal direction from data that are independent and identically distributed (i.i.d.), but little is understood about this problem in the more general relational setting with multiple types of interacting entities. This work examines the task of inferring the causal direction of peer dependence in relational data. We show that, in contrast to the i.i.d. setting, the direction of peer dependence can be inferred using simple procedures, regardless of the form of the underlying distribution, and we provide a theoretical characterization on the identifiability of direction. We then examine the conditions under which the presence of confounding can be detected. Finally, we demonstrate the efficacy of the proposed methods with synthetic experiments, and we provide an application on real-world data.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('771','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_771\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Inferring the direction of causal dependence from observational data is a fundamental problem in many scientific fields. Significant progress has been made in inferring causal direction from data that are independent and identically distributed (i.i.d.), but little is understood about this problem in the more general relational setting with multiple types of interacting entities. This work examines the task of inferring the causal direction of peer dependence in relational data. We show that, in contrast to the i.i.d. setting, the direction of peer dependence can be inferred using simple procedures, regardless of the form of the underlying distribution, and we provide a theoretical characterization on the identifiability of direction. We then examine the conditions under which the presence of confounding can be detected. Finally, we demonstrate the efficacy of the proposed methods with synthetic experiments, and we provide an application on real-world data.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('771','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_771\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/auai.org\/uai2016\/proceedings\/papers\/217.pdf\" title=\"http:\/\/auai.org\/uai2016\/proceedings\/papers\/217.pdf\" target=\"_blank\">http:\/\/auai.org\/uai2016\/proceedings\/papers\/217.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('771','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Shiri Dori-Hacohen, David Jensen, James Allan<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('772','tp_abstract')\" style=\"cursor:pointer;\">Controversy Detection in Wikipedia Using Collective Classification<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 39th International ACM SIGIR conference on \r\n Research and Development in Information Retrieval, SIGIR 2016, Pisa, \r\n Italy, July 17-21, 2016, <\/span><span class=\"tp_pub_additional_pages\">pp. 797\u2013800, <\/span><span class=\"tp_pub_additional_publisher\">ACM, <\/span><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_772\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('772','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_772\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('772','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_772\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('772','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_772\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/sigir\/Dori-HacohenJA16,<br \/>\r\ntitle = {Controversy Detection in Wikipedia Using Collective Classification},<br \/>\r\nauthor = {Shiri Dori-Hacohen and David Jensen and James Allan},<br \/>\r\nurl = {https:\/\/doi.org\/10.1145\/2911451.2914745},<br \/>\r\ndoi = {10.1145\/2911451.2914745},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\nbooktitle = {Proceedings of the 39th International ACM SIGIR conference on <br \/>\r\n Research and Development in Information Retrieval, SIGIR 2016, Pisa, <br \/>\r\n Italy, July 17-21, 2016},<br \/>\r\npages = {797--800},<br \/>\r\npublisher = {ACM},<br \/>\r\nabstract = {Concerns over personalization in IR have sparked an interest in detection and analysis of controversial topics. Accurate detection would enable many beneficial applications, such as alerting search users to controversy. Wikipedia's broad coverage and rich metadata offer a valuable resource for this problem. We hypothesize that intensities of controversy among related pages are not independent; thus, we propose a stacked model which exploits the dependencies among related pages. Our approach improves classification of controversial web pages when compared to a model that examines each page in isolation, demonstrating that controversial topics exhibit homophily. Using notions of similarity to construct a subnetwork for collective classification, rather than using the default network present in the relational data, leads to improved classification with wider applications for semi-structured datasets, with the effects most pronounced when a small set of neighbors is used.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('772','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_772\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Concerns over personalization in IR have sparked an interest in detection and analysis of controversial topics. Accurate detection would enable many beneficial applications, such as alerting search users to controversy. Wikipedia's broad coverage and rich metadata offer a valuable resource for this problem. We hypothesize that intensities of controversy among related pages are not independent; thus, we propose a stacked model which exploits the dependencies among related pages. Our approach improves classification of controversial web pages when compared to a model that examines each page in isolation, demonstrating that controversial topics exhibit homophily. Using notions of similarity to construct a subnetwork for collective classification, rather than using the default network present in the relational data, leads to improved classification with wider applications for semi-structured datasets, with the effects most pronounced when a small set of neighbors is used.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('772','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_772\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.1145\/2911451.2914745\" title=\"https:\/\/doi.org\/10.1145\/2911451.2914745\" target=\"_blank\">https:\/\/doi.org\/10.1145\/2911451.2914745<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1145\/2911451.2914745\" title=\"Follow DOI:10.1145\/2911451.2914745\" target=\"_blank\">doi:10.1145\/2911451.2914745<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('772','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Katerina Marazopoulou, Rumi Ghosh, Prasanth Lade, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('769','tp_abstract')\" style=\"cursor:pointer;\">Causal Discovery for Manufacturing Domains<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_769\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('769','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_769\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('769','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_769\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('769','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_769\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/MarazopoulouGLJ16,<br \/>\r\ntitle = {Causal Discovery for Manufacturing Domains},<br \/>\r\nauthor = {Katerina Marazopoulou and Rumi Ghosh and Prasanth Lade and David Jensen},<br \/>\r\nurl = {http:\/\/arxiv.org\/abs\/1605.04056},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/1605.04056},<br \/>\r\nabstract = {Yield and quality improvement is of paramount importance to any manufacturing company. One of the ways of improving yield is through discovery of the root causal factors affecting yield. We propose the use of data-driven interpretable causal models to identify key factors affecting yield. We focus on factors that are measured in different stages of production and testing in the manufacturing cycle of a product. We apply causal structure learning techniques on real data collected from this line. Specifically, the goal of this work is to learn interpretable causal models from observational data produced by manufacturing lines. Emphasis has been given to the interpretability of the models to make them actionable in the field of manufacturing. We highlight the challenges presented by assembly line data and propose ways to alleviate them.We also identify unique characteristics of data originating from assembly lines and how to leverage them in order to improve causal discovery. Standard evaluation techniques for causal structure learning shows that the learned causal models seem to closely represent the underlying latent causal relationship between different factors in the production process. These results were also validated by manufacturing domain experts who found them promising. This work demonstrates how data mining and knowledge discovery can be used for root cause analysis in the domain of manufacturing and connected industry.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('769','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_769\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Yield and quality improvement is of paramount importance to any manufacturing company. One of the ways of improving yield is through discovery of the root causal factors affecting yield. We propose the use of data-driven interpretable causal models to identify key factors affecting yield. We focus on factors that are measured in different stages of production and testing in the manufacturing cycle of a product. We apply causal structure learning techniques on real data collected from this line. Specifically, the goal of this work is to learn interpretable causal models from observational data produced by manufacturing lines. Emphasis has been given to the interpretability of the models to make them actionable in the field of manufacturing. We highlight the challenges presented by assembly line data and propose ways to alleviate them.We also identify unique characteristics of data originating from assembly lines and how to leverage them in order to improve causal discovery. Standard evaluation techniques for causal structure learning shows that the learned causal models seem to closely represent the underlying latent causal relationship between different factors in the production process. These results were also validated by manufacturing domain experts who found them promising. This work demonstrates how data mining and knowledge discovery can be used for root cause analysis in the domain of manufacturing and connected industry.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('769','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_769\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/arxiv.org\/abs\/1605.04056\" title=\"http:\/\/arxiv.org\/abs\/1605.04056\" target=\"_blank\">http:\/\/arxiv.org\/abs\/1605.04056<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('769','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Dan Garant, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('723','tp_abstract')\" style=\"cursor:pointer;\">Evaluating causal models by comparing interventional distributions<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2016<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_723\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('723','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_723\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('723','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_723\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('723','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_723\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{garant2016evaluating,<br \/>\r\ntitle = {Evaluating causal models by comparing interventional distributions},<br \/>\r\nauthor = {Dan Garant and David Jensen},<br \/>\r\nurl = {https:\/\/arxiv.org\/abs\/1608.04698},<br \/>\r\nyear  = {2016},<br \/>\r\ndate = {2016-01-01},<br \/>\r\njournal = {arXiv preprint arXiv:1608.04698},<br \/>\r\nabstract = {The predominant method for evaluating the quality of causal models is to measure the graphical accuracy of the learned model structure. We present an alternative method for evaluating causal models that directly measures the accuracy of estimated interventional distributions. We contrast such distributional measures with structural measures, such as structural Hamming distance and structural intervention distance, showing that structural measures often correspond poorly to the accuracy of estimated interventional distributions. We use a number of real and synthetic datasets to illustrate various scenarios in which structural measures provide misleading results with respect to algorithm selection and parameter tuning, and we recommend that distributional measures become the new standard for evaluating causal models.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('723','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_723\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The predominant method for evaluating the quality of causal models is to measure the graphical accuracy of the learned model structure. We present an alternative method for evaluating causal models that directly measures the accuracy of estimated interventional distributions. We contrast such distributional measures with structural measures, such as structural Hamming distance and structural intervention distance, showing that structural measures often correspond poorly to the accuracy of estimated interventional distributions. We use a number of real and synthetic datasets to illustrate various scenarios in which structural measures provide misleading results with respect to algorithm selection and parameter tuning, and we recommend that distributional measures become the new standard for evaluating causal models.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('723','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_723\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/arxiv.org\/abs\/1608.04698\" title=\"https:\/\/arxiv.org\/abs\/1608.04698\" target=\"_blank\">https:\/\/arxiv.org\/abs\/1608.04698<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('723','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2015\">2015<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Phillip B. Kirlin, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('765','tp_abstract')\" style=\"cursor:pointer;\">Learning to Uncover Deep Musical Structure<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, \r\n January 25-30, 2015, Austin, Texas, USA, <\/span><span class=\"tp_pub_additional_pages\">pp. 1770\u20131776, <\/span><span class=\"tp_pub_additional_publisher\">AAAI Press, <\/span><span class=\"tp_pub_additional_year\">2015<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_765\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('765','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_765\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('765','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_765\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('765','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_765\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/aaai\/KirlinJ15,<br \/>\r\ntitle = {Learning to Uncover Deep Musical Structure},<br \/>\r\nauthor = {Phillip B. Kirlin and David Jensen},<br \/>\r\nurl = {http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI15\/paper\/view\/9757},<br \/>\r\nyear  = {2015},<br \/>\r\ndate = {2015-01-01},<br \/>\r\nbooktitle = {Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, <br \/>\r\n January 25-30, 2015, Austin, Texas, USA},<br \/>\r\npages = {1770--1776},<br \/>\r\npublisher = {AAAI Press},<br \/>\r\nabstract = {The overarching goal of music theory is to explain the inner workings of a musical composition by examining the structure of the composition. Schenkerian music theory supposes that Western tonal compositions can be viewed as hierarchies of musical objects. The process of Schenkerian analysis reveals this hierarchy by identifying connections between notes or chords of a composition that illustrate both the small- and large-scale construction of the music. We present a new probabilistic model of this variety of music analysis, details of how the parameters of the model can be learned from a corpus, an algorithm for deriving the most probable analysis for a given piece of music, and both quantitative and human-based evaluations of the algorithm's performance. This represents the first large-scale data-driven computational approach to hierarchical music analysis.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('765','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_765\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The overarching goal of music theory is to explain the inner workings of a musical composition by examining the structure of the composition. Schenkerian music theory supposes that Western tonal compositions can be viewed as hierarchies of musical objects. The process of Schenkerian analysis reveals this hierarchy by identifying connections between notes or chords of a composition that illustrate both the small- and large-scale construction of the music. We present a new probabilistic model of this variety of music analysis, details of how the parameters of the model can be learned from a corpus, an algorithm for deriving the most probable analysis for a given piece of music, and both quantitative and human-based evaluations of the algorithm's performance. This represents the first large-scale data-driven computational approach to hierarchical music analysis.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('765','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_765\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI15\/paper\/view\/9757\" title=\"http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI15\/paper\/view\/9757\" target=\"_blank\">http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI15\/paper\/view\/9757<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('765','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Jerod J. Weinman, David Jensen, David Lopatto<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('766','tp_abstract')\" style=\"cursor:pointer;\">Teaching Computing as Science in a Research Experience<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 46th ACM Technical Symposium on Computer Science \r\n Education, SIGCSE 2015, Kansas City, MO, USA, March 4-7, 2015, <\/span><span class=\"tp_pub_additional_pages\">pp. 24\u201329, <\/span><span class=\"tp_pub_additional_publisher\">ACM, <\/span><span class=\"tp_pub_additional_year\">2015<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_766\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('766','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_766\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('766','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_766\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('766','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_766\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/sigcse\/WeinmanJL15,<br \/>\r\ntitle = {Teaching Computing as Science in a Research Experience},<br \/>\r\nauthor = {Jerod J. Weinman and David Jensen and David Lopatto},<br \/>\r\nurl = {https:\/\/doi.org\/10.1145\/2676723.2677231},<br \/>\r\ndoi = {10.1145\/2676723.2677231},<br \/>\r\nyear  = {2015},<br \/>\r\ndate = {2015-01-01},<br \/>\r\nbooktitle = {Proceedings of the 46th ACM Technical Symposium on Computer Science <br \/>\r\n Education, SIGCSE 2015, Kansas City, MO, USA, March 4-7, 2015},<br \/>\r\npages = {24--29},<br \/>\r\npublisher = {ACM},<br \/>\r\nabstract = {Many instructors and institutions offer research experiences and training in computing research methods. However, in a national survey, we find that undergraduate students rate their computing research experiences lower than students in other STEM fields. To address this learning gap, we have offered summer undergraduate research experiences in computing that include not only instruction in the important mechanics of research but also grounding in a philosophy of computing science that emphasizes generalized explanation of behavior as a means for control and prediction. After five years, survey results indicate the experience helps close the gap between CS and other STEM fields in benefits gained.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('766','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_766\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Many instructors and institutions offer research experiences and training in computing research methods. However, in a national survey, we find that undergraduate students rate their computing research experiences lower than students in other STEM fields. To address this learning gap, we have offered summer undergraduate research experiences in computing that include not only instruction in the important mechanics of research but also grounding in a philosophy of computing science that emphasizes generalized explanation of behavior as a means for control and prediction. After five years, survey results indicate the experience helps close the gap between CS and other STEM fields in benefits gained.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('766','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_766\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.1145\/2676723.2677231\" title=\"https:\/\/doi.org\/10.1145\/2676723.2677231\" target=\"_blank\">https:\/\/doi.org\/10.1145\/2676723.2677231<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1145\/2676723.2677231\" title=\"Follow DOI:10.1145\/2676723.2677231\" target=\"_blank\">doi:10.1145\/2676723.2677231<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('766','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Katerina Marazopoulou, Marc Maier, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('767','tp_abstract')\" style=\"cursor:pointer;\">Learning the Structure of Causal Models with Relational and Temporal \r\n Dependence<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the UAI 2015 Workshop on Advances in Causal Inference \r\n co-located with the 31st Conference on Uncertainty in Artificial Intelligence \r\n (UAI 2015), Amsterdam, The Netherlands, July 16, 2015, <\/span><span class=\"tp_pub_additional_pages\">pp. 66\u201375, <\/span><span class=\"tp_pub_additional_publisher\">CEUR-WS.org, <\/span><span class=\"tp_pub_additional_year\">2015<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_767\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('767','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_767\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('767','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_767\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('767','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_767\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/uai\/MarazopoulouMJ15,<br \/>\r\ntitle = {Learning the Structure of Causal Models with Relational and Temporal <br \/>\r\n Dependence},<br \/>\r\nauthor = {Katerina Marazopoulou and Marc Maier and David Jensen},<br \/>\r\nurl = {http:\/\/ceur-ws.org\/Vol-1504\/uai2015aci_paper6.pdf},<br \/>\r\nyear  = {2015},<br \/>\r\ndate = {2015-01-01},<br \/>\r\nbooktitle = {Proceedings of the UAI 2015 Workshop on Advances in Causal Inference <br \/>\r\n co-located with the 31st Conference on Uncertainty in Artificial Intelligence <br \/>\r\n (UAI 2015), Amsterdam, The Netherlands, July 16, 2015},<br \/>\r\nvolume = {1504},<br \/>\r\npages = {66--75},<br \/>\r\npublisher = {CEUR-WS.org},<br \/>\r\nseries = {CEUR Workshop Proceedings},<br \/>\r\nabstract = {Many real-world domains are inherently relational and temporal\u2014they consist of heterogeneous entities that interact with each other over time. Effective reasoning about causality in such domains requires representations that explicitly model relational and temporal dependence. In this work, we provide a formalization of temporal relational models. We define temporal extensions to abstract ground graphs\u2014a lifted representation that abstracts paths of dependence over all possible ground graphs. Temporal abstract ground graphs enable a sound and complete method for answering d-separation queries on temporal relational models. These methods provide the foundation for a constraint-based algorithm, TRCD, that learns causal models from temporal relational data. We provide experimental evidence that demonstrates the need to explicitly represent time when inferring causal dependence. We also demonstrate the expressive gain of TRCD compared to earlier algorithms that do not explicitly represent time.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('767','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_767\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Many real-world domains are inherently relational and temporal\u2014they consist of heterogeneous entities that interact with each other over time. Effective reasoning about causality in such domains requires representations that explicitly model relational and temporal dependence. In this work, we provide a formalization of temporal relational models. We define temporal extensions to abstract ground graphs\u2014a lifted representation that abstracts paths of dependence over all possible ground graphs. Temporal abstract ground graphs enable a sound and complete method for answering d-separation queries on temporal relational models. These methods provide the foundation for a constraint-based algorithm, TRCD, that learns causal models from temporal relational data. We provide experimental evidence that demonstrates the need to explicitly represent time when inferring causal dependence. We also demonstrate the expressive gain of TRCD compared to earlier algorithms that do not explicitly represent time.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('767','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_767\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ceur-ws.org\/Vol-1504\/uai2015aci_paper6.pdf\" title=\"http:\/\/ceur-ws.org\/Vol-1504\/uai2015aci_paper6.pdf\" target=\"_blank\">http:\/\/ceur-ws.org\/Vol-1504\/uai2015aci_paper6.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('767','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2014\">2014<\/h3><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Lisa Friedland, Amanda Gentzel, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('761','tp_abstract')\" style=\"cursor:pointer;\">Classifier-adjusted density estimation for anomaly detection and one-class classification<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 2014 SIAM International Conference on Data Mining, <\/span><span class=\"tp_pub_additional_pages\">pp. 578\u2013586, <\/span><span class=\"tp_pub_additional_organization\">Society for Industrial and Applied Mathematics <\/span><span class=\"tp_pub_additional_year\">2014<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_761\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('761','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_761\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('761','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_761\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('761','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_761\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{friedland2014classifier,<br \/>\r\ntitle = {Classifier-adjusted density estimation for anomaly detection and one-class classification},<br \/>\r\nauthor = {Lisa Friedland and Amanda Gentzel and David Jensen},<br \/>\r\nurl = {https:\/\/epubs.siam.org\/doi\/pdf\/10.1137\/1.9781611973440.67},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-01-01},<br \/>\r\nbooktitle = {Proceedings of the 2014 SIAM International Conference on Data Mining},<br \/>\r\npages = {578--586},<br \/>\r\norganization = {Society for Industrial and Applied Mathematics},<br \/>\r\nabstract = {Density estimation methods are often regarded as unsuitable for anomaly detection in high-dimensional data due to the difficulty of estimating multivariate probability distributions. Instead, the scores from popular distance- and local-density-based methods, such as local outlier factor (LOF), are used as surrogates for probability densities. We question this infeasibility assumption and explore a family of simple statistically-based density estimates constructed by combining a probabilistic classifier with a naive density estimate. Across a number of semi-supervised and unsupervised problems formed from real-world data sets, we show that these methods are competitive with LOF and that even simple density estimates that assume attribute independence can perform strongly. We show that these density estimation methods scale well to data with high dimensionality and that they are robust to the problem of irrelevant attributes that plagues methods based on local estimates.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('761','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_761\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Density estimation methods are often regarded as unsuitable for anomaly detection in high-dimensional data due to the difficulty of estimating multivariate probability distributions. Instead, the scores from popular distance- and local-density-based methods, such as local outlier factor (LOF), are used as surrogates for probability densities. We question this infeasibility assumption and explore a family of simple statistically-based density estimates constructed by combining a probabilistic classifier with a naive density estimate. Across a number of semi-supervised and unsupervised problems formed from real-world data sets, we show that these methods are competitive with LOF and that even simple density estimates that assume attribute independence can perform strongly. We show that these density estimation methods scale well to data with high dimensionality and that they are robust to the problem of irrelevant attributes that plagues methods based on local estimates.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('761','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_761\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/epubs.siam.org\/doi\/pdf\/10.1137\/1.9781611973440.67\" title=\"https:\/\/epubs.siam.org\/doi\/pdf\/10.1137\/1.9781611973440.67\" target=\"_blank\">https:\/\/epubs.siam.org\/doi\/pdf\/10.1137\/1.9781611973440.67<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('761','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Arbour, Katerina Marazopoulou, Dan Garant, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('763','tp_abstract')\" style=\"cursor:pointer;\">Propensity Score Matching for Causal Inference with Relational Data<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the UAI 2014 Workshop Causal Inference: Learning \r\n and Prediction co-located with 30th Conference on Uncertainty in Artificial \r\n Intelligence (UAI 2014), Quebec City, Canada, July 27, 2014, <\/span><span class=\"tp_pub_additional_pages\">pp. 25\u201334, <\/span><span class=\"tp_pub_additional_publisher\">CEUR-WS.org, <\/span><span class=\"tp_pub_additional_year\">2014<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_763\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('763','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_763\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('763','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_763\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('763','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_763\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/uai\/ArbourMGJ14,<br \/>\r\ntitle = {Propensity Score Matching for Causal Inference with Relational Data},<br \/>\r\nauthor = {David Arbour and Katerina Marazopoulou and Dan Garant and David Jensen},<br \/>\r\nurl = {http:\/\/ceur-ws.org\/Vol-1274\/uai2014ci_paper5.pdf},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-01-01},<br \/>\r\nbooktitle = {Proceedings of the UAI 2014 Workshop Causal Inference: Learning <br \/>\r\n and Prediction co-located with 30th Conference on Uncertainty in Artificial <br \/>\r\n Intelligence (UAI 2014), Quebec City, Canada, July 27, 2014},<br \/>\r\nvolume = {1274},<br \/>\r\npages = {25--34},<br \/>\r\npublisher = {CEUR-WS.org},<br \/>\r\nseries = {CEUR Workshop Proceedings},<br \/>\r\nabstract = {Propensity score matching (PSM) is a widely used method for performing causal inference with observational data. PSM requires fully specifying the set of confounding variables of treatment and outcome. In the case of relational data, this set may include non-intuitive relational variables, i.e., variables derived from the relational structure of the data. In this work, we provide an automated method to derive these relational variables based on the relational structure and a set of naive confounders. This automatic construction includes two unusual classes of variables: relational degree and entity identifiers. We provide experimental evidence that demonstrates the utility of these variables in accounting for certain latent confounders. Finally, through a set of synthetic experiments, we show that our method improves the performance of PSM for causal inference with relational data.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('763','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_763\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Propensity score matching (PSM) is a widely used method for performing causal inference with observational data. PSM requires fully specifying the set of confounding variables of treatment and outcome. In the case of relational data, this set may include non-intuitive relational variables, i.e., variables derived from the relational structure of the data. In this work, we provide an automated method to derive these relational variables based on the relational structure and a set of naive confounders. This automatic construction includes two unusual classes of variables: relational degree and entity identifiers. We provide experimental evidence that demonstrates the utility of these variables in accounting for certain latent confounders. Finally, through a set of synthetic experiments, we show that our method improves the performance of PSM for causal inference with relational data.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('763','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_763\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ceur-ws.org\/Vol-1274\/uai2014ci_paper5.pdf\" title=\"http:\/\/ceur-ws.org\/Vol-1274\/uai2014ci_paper5.pdf\" target=\"_blank\">http:\/\/ceur-ws.org\/Vol-1274\/uai2014ci_paper5.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('763','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Katerina Marazopoulou, David Arbour, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('764','tp_abstract')\" style=\"cursor:pointer;\">Refining the Semantics of Social Influence<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2014<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_764\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('764','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_764\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('764','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_764\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('764','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_764\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/MarazopoulouAJ14,<br \/>\r\ntitle = {Refining the Semantics of Social Influence},<br \/>\r\nauthor = {Katerina Marazopoulou and David Arbour and David Jensen},<br \/>\r\nurl = {http:\/\/arxiv.org\/abs\/1412.5238},<br \/>\r\nyear  = {2014},<br \/>\r\ndate = {2014-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/1412.5238},<br \/>\r\nabstract = {With the proliferation of network data, researchers are increasingly focusing on questions investigating phenomena occurring on networks. This often includes analysis of peer-effects, i.e., how the connections of an individual affect that individual's behavior. This type of influence is not limited to direct connections of an individual (such as friends), but also to individuals that are connected through longer paths (for example, friends of friends, or friends of friends of friends). In this work, we identify an ambiguity in the definition of what constitutes the extended neighborhood of an individual. This ambiguity gives rise to different semantics and supports different types of underlying phenomena. We present experimental results, both on synthetic and real networks, that quantify differences among the sets of extended neighbors under different semantics. Finally, we provide experimental evidence that demonstrates how the use of different semantics affects model selection.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('764','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_764\" style=\"display:none;\"><div class=\"tp_abstract_entry\">With the proliferation of network data, researchers are increasingly focusing on questions investigating phenomena occurring on networks. This often includes analysis of peer-effects, i.e., how the connections of an individual affect that individual's behavior. This type of influence is not limited to direct connections of an individual (such as friends), but also to individuals that are connected through longer paths (for example, friends of friends, or friends of friends of friends). In this work, we identify an ambiguity in the definition of what constitutes the extended neighborhood of an individual. This ambiguity gives rise to different semantics and supports different types of underlying phenomena. We present experimental results, both on synthetic and real networks, that quantify differences among the sets of extended neighbors under different semantics. Finally, we provide experimental evidence that demonstrates how the use of different semantics affects model selection.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('764','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_764\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/arxiv.org\/abs\/1412.5238\" title=\"http:\/\/arxiv.org\/abs\/1412.5238\" target=\"_blank\">http:\/\/arxiv.org\/abs\/1412.5238<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('764','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2013\">2013<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">David Arbour, James Atwood, Ahmed El-Kishky, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('760','tp_abstract')\" style=\"cursor:pointer;\">Agglomerative Clustering of Bagged Data Using Joint Distributions<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_year\">2013<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_760\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('760','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_760\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('760','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_760\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('760','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_760\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{arbour2013agglomerative,<br \/>\r\ntitle = {Agglomerative Clustering of Bagged Data Using Joint Distributions},<br \/>\r\nauthor = {David Arbour and James Atwood and Ahmed El-Kishky and David Jensen},<br \/>\r\nurl = {http:\/\/citeseerx.ist.psu.edu\/viewdoc\/download?doi=10.1.1.703.7198&rep=rep1&type=pdf},<br \/>\r\nyear  = {2013},<br \/>\r\ndate = {2013-01-01},<br \/>\r\npublisher = {Citeseer},<br \/>\r\nabstract = {Current methods for hierarchical clustering of data either operate on features of the data or make limiting model assumptions. We present the hierarchy discovery algorithm (HDA), a model-based hierarchical clustering method based on explicit comparison of joint distributions via Bayesian network learning for predefined groups of data. HDA works on both continuous and discrete data and offers a model-based approach to agglomerative clustering that does not require prespecification of the model dependency structure.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('760','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_760\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Current methods for hierarchical clustering of data either operate on features of the data or make limiting model assumptions. We present the hierarchy discovery algorithm (HDA), a model-based hierarchical clustering method based on explicit comparison of joint distributions via Bayesian network learning for predefined groups of data. HDA works on both continuous and discrete data and offers a model-based approach to agglomerative clustering that does not require prespecification of the model dependency structure.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('760','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_760\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/citeseerx.ist.psu.edu\/viewdoc\/download?doi=10.1.1.703.7198&amp;rep=rep1&amp;type=pdf\" title=\"http:\/\/citeseerx.ist.psu.edu\/viewdoc\/download?doi=10.1.1.703.7198&amp;rep=rep1&a[...]\" target=\"_blank\">http:\/\/citeseerx.ist.psu.edu\/viewdoc\/download?doi=10.1.1.703.7198&amp;rep=rep1&a[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('760','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Lisa Friedland, David Jensen, Michael Lavine<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('758','tp_abstract')\" style=\"cursor:pointer;\">Copy or Coincidence? A Model for Detecting Social Influence and \r\n Duplication Events<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 30th International Conference on Machine Learning, \r\n ICML 2013, Atlanta, GA, USA, 16-21 June 2013, <\/span><span class=\"tp_pub_additional_pages\">pp. 1175\u20131183, <\/span><span class=\"tp_pub_additional_publisher\">JMLR.org, <\/span><span class=\"tp_pub_additional_year\">2013<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_758\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('758','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_758\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('758','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_758\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('758','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_758\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/icml\/FriedlandJL13,<br \/>\r\ntitle = {Copy or Coincidence? A Model for Detecting Social Influence and <br \/>\r\n Duplication Events},<br \/>\r\nauthor = {Lisa Friedland and David Jensen and Michael Lavine},<br \/>\r\nurl = {http:\/\/proceedings.mlr.press\/v28\/friedland13.html},<br \/>\r\nyear  = {2013},<br \/>\r\ndate = {2013-01-01},<br \/>\r\nbooktitle = {Proceedings of the 30th International Conference on Machine Learning, <br \/>\r\n ICML 2013, Atlanta, GA, USA, 16-21 June 2013},<br \/>\r\nvolume = {28},<br \/>\r\npages = {1175--1183},<br \/>\r\npublisher = {JMLR.org},<br \/>\r\nseries = {JMLR Workshop and Conference Proceedings},<br \/>\r\nabstract = {In this paper, we analyze the task of inferring rare links between pairs of entities that seem too similar to have occurred by chance. Variations of this task appear in such diverse areas as social network analysis, security, fraud detection, and entity resolution. To address the task in a general form, we propose a simple, flexible mixture model in which most entities are generated independently from a distribution but a small number of pairs are constrained to be similar. We predict the true pairs using a likelihood ratio that trades off the entities\u2019 similarity with their rarity. This method always outperforms using only similarity; however, with certain parameter settings, similarity turns out to be surprisingly competitive. Using real data, we apply the model to detect twins given their birth weights and to re-identify cell phone users based on distinctive usage patterns.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('758','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_758\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In this paper, we analyze the task of inferring rare links between pairs of entities that seem too similar to have occurred by chance. Variations of this task appear in such diverse areas as social network analysis, security, fraud detection, and entity resolution. To address the task in a general form, we propose a simple, flexible mixture model in which most entities are generated independently from a distribution but a small number of pairs are constrained to be similar. We predict the true pairs using a likelihood ratio that trades off the entities\u2019 similarity with their rarity. This method always outperforms using only similarity; however, with certain parameter settings, similarity turns out to be surprisingly competitive. Using real data, we apply the model to detect twins given their birth weights and to re-identify cell phone users based on distinctive usage patterns.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('758','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_758\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/proceedings.mlr.press\/v28\/friedland13.html\" title=\"http:\/\/proceedings.mlr.press\/v28\/friedland13.html\" target=\"_blank\">http:\/\/proceedings.mlr.press\/v28\/friedland13.html<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('758','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Marc Maier, Katerina Marazopoulou, David Arbour, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('759','tp_abstract')\" style=\"cursor:pointer;\">A Sound and Complete Algorithm for Learning Causal Models from Relational \r\n Data<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial \r\n Intelligence, UAI 2013, Bellevue, WA, USA, August 11-15, 2013, <\/span><span class=\"tp_pub_additional_publisher\">AUAI Press, <\/span><span class=\"tp_pub_additional_year\">2013<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_759\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('759','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_759\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('759','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_759\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('759','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_759\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{DBLP:conf\/uai\/MaierMAJ13,<br \/>\r\ntitle = {A Sound and Complete Algorithm for Learning Causal Models from Relational <br \/>\r\n Data},<br \/>\r\nauthor = {Marc Maier and Katerina Marazopoulou and David Arbour and David Jensen},<br \/>\r\nurl = {https:\/\/dslpitt.org\/uai\/displayArticleDetails.jsp?mmnu=1&smnu=2&article_id=2398&proceeding_id=29},<br \/>\r\nyear  = {2013},<br \/>\r\ndate = {2013-01-01},<br \/>\r\nbooktitle = {Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial <br \/>\r\n Intelligence, UAI 2013, Bellevue, WA, USA, August 11-15, 2013},<br \/>\r\npublisher = {AUAI Press},<br \/>\r\nabstract = {In this paper, we analyze the task of inferring rare links between pairs of entities that seem too similar to have occurred by chance. Variations of this task appear in such diverse areas as social network analysis, security, fraud detection, and entity resolution. To address the task in a general form, we propose a simple, flexible mixture model in which most entities are generated independently from a distribution but a small number of pairs are constrained to be similar. We predict the true pairs using a likelihood ratio that trades off the entities\u2019 similarity with their rarity. This method always outperforms using only similarity; however, with certain parameter settings, similarity turns out to be surprisingly competitive. Using real data, we apply the model to detect twins given their birth weights and to re-identify cell phone users based on distinctive usage patterns.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('759','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_759\" style=\"display:none;\"><div class=\"tp_abstract_entry\">In this paper, we analyze the task of inferring rare links between pairs of entities that seem too similar to have occurred by chance. Variations of this task appear in such diverse areas as social network analysis, security, fraud detection, and entity resolution. To address the task in a general form, we propose a simple, flexible mixture model in which most entities are generated independently from a distribution but a small number of pairs are constrained to be similar. We predict the true pairs using a likelihood ratio that trades off the entities\u2019 similarity with their rarity. This method always outperforms using only similarity; however, with certain parameter settings, similarity turns out to be surprisingly competitive. Using real data, we apply the model to detect twins given their birth weights and to re-identify cell phone users based on distinctive usage patterns.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('759','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_759\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dslpitt.org\/uai\/displayArticleDetails.jsp?mmnu=1&amp;smnu=2&amp;article_id=2398&amp;proceeding_id=29\" title=\"https:\/\/dslpitt.org\/uai\/displayArticleDetails.jsp?mmnu=1&amp;smnu=2&amp;article_[...]\" target=\"_blank\">https:\/\/dslpitt.org\/uai\/displayArticleDetails.jsp?mmnu=1&amp;smnu=2&amp;article_[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('759','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Marc Maier, Katerina Marazopoulou, David Arbour, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('762','tp_abstract')\" style=\"cursor:pointer;\">Flattening network data for causal discovery: What could go wrong?<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Workshop on Information in Networks, <\/span><span class=\"tp_pub_additional_year\">2013<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_762\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('762','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_762\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('762','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_762\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('762','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_762\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{maier2013flattening,<br \/>\r\ntitle = {Flattening network data for causal discovery: What could go wrong?},<br \/>\r\nauthor = {Marc Maier and Katerina Marazopoulou and David Arbour and David Jensen},<br \/>\r\nurl = {https:\/\/www.semanticscholar.org\/paper\/Flattening-network-data-for-causal-discovery-%3A-What-Maier-Marazopoulou\/c327100636c022c259f5e1bf2d7fcbbd0b048935},<br \/>\r\nyear  = {2013},<br \/>\r\ndate = {2013-01-01},<br \/>\r\nbooktitle = {Workshop on Information in Networks},<br \/>\r\nvolume = {64},<br \/>\r\nabstract = {Methods for learning causal dependencies from observational data have been the focus of decades of work in social science, statistics, machine learning, and philosophy [9, 10, 11]. Much of the theoretical and practical work on causal discovery has focused on propositional representations. Propositional models effectively represent individual directed causal dependencies (e.g., path analysis, Bayesian networks) or conditional distributions of some outcome variable (e.g., linear regression, decision trees). However, propositional representations are limited to modeling independent and identically distributed (IID) data of a single entity type. Many real-world systems involve heterogeneous, interacting entities with probabilistic dependencies that cross the boundaries of those entities (i.e., non-IID data with multiple entity types and relationships). These systems produce network, or relational, data, and they are of paramount interest to researchers and practitioners across a wide range of disciplines. To model such data, researchers in statistics and computer science have devised more expressive classes of directed graphical models, such as probabilistic relational models (PRMs) [2] and directed acyclic probabilistic entityrelationship (DAPER) models [4]. Despite the assumptions embedded in propositional models, a common practice is to flatten, or propositionalize, relational data and use existing algorithms [5] (see Figure 1, focusing on algorithms that learn causal graphical models). While there are statistical concerns, this process is generally innocuous if the task is to model statistical associations for predictive inference. In contrast, to learn causal structure, estimate causal effects, or support inference over interventions, the effects of flattening inherently relational data can be particularly deleterious. In this paper, we identify four classes of potential issues that can occur with a propositionalization strategy as opposed to embracing a more expressive representation that would not succumb to these problems. We also present empirical results comparing the effectiveness of two theoretically sound and complete algorithms that learn causal structure: PC\u2014a widely used constraint-based, propositional algorithm for causal discovery [11], and RCD\u2014a recently developed constraint-based algorithm that reasons over a relational representation [6].},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('762','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_762\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Methods for learning causal dependencies from observational data have been the focus of decades of work in social science, statistics, machine learning, and philosophy [9, 10, 11]. Much of the theoretical and practical work on causal discovery has focused on propositional representations. Propositional models effectively represent individual directed causal dependencies (e.g., path analysis, Bayesian networks) or conditional distributions of some outcome variable (e.g., linear regression, decision trees). However, propositional representations are limited to modeling independent and identically distributed (IID) data of a single entity type. Many real-world systems involve heterogeneous, interacting entities with probabilistic dependencies that cross the boundaries of those entities (i.e., non-IID data with multiple entity types and relationships). These systems produce network, or relational, data, and they are of paramount interest to researchers and practitioners across a wide range of disciplines. To model such data, researchers in statistics and computer science have devised more expressive classes of directed graphical models, such as probabilistic relational models (PRMs) [2] and directed acyclic probabilistic entityrelationship (DAPER) models [4]. Despite the assumptions embedded in propositional models, a common practice is to flatten, or propositionalize, relational data and use existing algorithms [5] (see Figure 1, focusing on algorithms that learn causal graphical models). While there are statistical concerns, this process is generally innocuous if the task is to model statistical associations for predictive inference. In contrast, to learn causal structure, estimate causal effects, or support inference over interventions, the effects of flattening inherently relational data can be particularly deleterious. In this paper, we identify four classes of potential issues that can occur with a propositionalization strategy as opposed to embracing a more expressive representation that would not succumb to these problems. We also present empirical results comparing the effectiveness of two theoretically sound and complete algorithms that learn causal structure: PC\u2014a widely used constraint-based, propositional algorithm for causal discovery [11], and RCD\u2014a recently developed constraint-based algorithm that reasons over a relational representation [6].<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('762','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_762\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/www.semanticscholar.org\/paper\/Flattening-network-data-for-causal-discovery-%3A-What-Maier-Marazopoulou\/c327100636c022c259f5e1bf2d7fcbbd0b048935\" title=\"https:\/\/www.semanticscholar.org\/paper\/Flattening-network-data-for-causal-discove[...]\" target=\"_blank\">https:\/\/www.semanticscholar.org\/paper\/Flattening-network-data-for-causal-discove[...]<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('762','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_misc\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Marc Maier, Katerina Marazopoulou, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('757','tp_abstract')\" style=\"cursor:pointer;\">Reasoning about Independence in Probabilistic Models of Relational \r\n Data<\/a> <span class=\"tp_pub_type tp_  misc\">Miscellaneous<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2013<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_757\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('757','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_757\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('757','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_757\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('757','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_757\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@misc{DBLP:journals\/corr\/abs-1302-4381,<br \/>\r\ntitle = {Reasoning about Independence in Probabilistic Models of Relational <br \/>\r\n Data},<br \/>\r\nauthor = {Marc Maier and Katerina Marazopoulou and David Jensen},<br \/>\r\nurl = {http:\/\/arxiv.org\/abs\/1302.4381},<br \/>\r\nyear  = {2013},<br \/>\r\ndate = {2013-01-01},<br \/>\r\njournal = {CoRR},<br \/>\r\nvolume = {abs\/1302.4381},<br \/>\r\nabstract = {We extend the theory of d-separation to cases in which data instances are not independent and identically distributed. We show that applying the rules of d-separation directly to the structure of probabilistic models of relational data inaccurately infers conditional independence. We introduce relational d-separation, a theory for deriving conditional independence facts from relational models. We provide a new representation, the abstract ground graph, that enables a sound, complete, and computationally efficient method for answering d-separation queries about relational models, and we present empirical results that demonstrate effectiveness.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {misc}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('757','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_757\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We extend the theory of d-separation to cases in which data instances are not independent and identically distributed. We show that applying the rules of d-separation directly to the structure of probabilistic models of relational data inaccurately infers conditional independence. We introduce relational d-separation, a theory for deriving conditional independence facts from relational models. We provide a new representation, the abstract ground graph, that enables a sound, complete, and computationally efficient method for answering d-separation queries about relational models, and we present empirical results that demonstrate effectiveness.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('757','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_757\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"ai ai-arxiv\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/arxiv.org\/abs\/1302.4381\" title=\"http:\/\/arxiv.org\/abs\/1302.4381\" target=\"_blank\">http:\/\/arxiv.org\/abs\/1302.4381<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('757','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2012\">2012<\/h3><div class=\"tp_publication tp_publication_phdthesis\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Matthew Rattigan<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('773','tp_abstract')\" style=\"cursor:pointer;\">Leveraging Relational Representations for Causal Discovery<\/a> <span class=\"tp_pub_type tp_  phdthesis\">PhD Thesis<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_year\">2012<\/span>, <span class=\"tp_pub_additional_isbn\">ISBN: 9781267786821<\/span><span class=\"tp_pub_additional_note\">, (AAI3545976)<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_773\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('773','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_773\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('773','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_773\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@phdthesis{10.5555\/2520420,<br \/>\r\ntitle = {Leveraging Relational Representations for Causal Discovery},<br \/>\r\nauthor = {Matthew Rattigan},<br \/>\r\nisbn = {9781267786821},<br \/>\r\nyear  = {2012},<br \/>\r\ndate = {2012-01-01},<br \/>\r\npublisher = {University of Massachusetts Amherst},<br \/>\r\nabstract = {This thesis represents a synthesis of relational learning and causal discovery, two subjects at the frontier of machine learning research. Relational learning investigates algorithms for constructing statistical models of data drawn from of multiple types of interrelated entities, and causal discovery investigates algorithms for constructing causal models from observational data. My work demonstrates that there exists a natural, methodological synergy between these two areas of study, and that despite the sometimes onerous nature of each, their combination (perhaps counterintuitively) can provide advances in the state of the art for both. Traditionally, propositional (or \"flat\") data representations have dominated the statistical sciences. These representations assume that data consist of independent and identically distributed (iid) entities which can be represented by a single data table. More recently, data scientists have increasingly focused on \"relational\" data sets that consist of interrelated, heterogeneous entities. However, relational learning and causal discovery are rarely combined. Relational representations are wholly absent from the literature where causality is discussed explicitly. Instead, the literature on causality that uses the framework of graphical models assumes that data are independent and identically distributed. This unexplored topical intersection represents an opportunity for advancement \u2014 by combining relational learning with causal reasoning, we can provide insight into the challenges found in each subject area. By adopting a causal viewpoint, we can clarify the mechanisms that produce previously identified pathologies in relational learning. Analogously, we can utilize relational data to establish and strengthen causal claims in ways that are impossible using only propositional representations.},<br \/>\r\nnote = {AAI3545976},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {phdthesis}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('773','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_773\" style=\"display:none;\"><div class=\"tp_abstract_entry\">This thesis represents a synthesis of relational learning and causal discovery, two subjects at the frontier of machine learning research. Relational learning investigates algorithms for constructing statistical models of data drawn from of multiple types of interrelated entities, and causal discovery investigates algorithms for constructing causal models from observational data. My work demonstrates that there exists a natural, methodological synergy between these two areas of study, and that despite the sometimes onerous nature of each, their combination (perhaps counterintuitively) can provide advances in the state of the art for both. Traditionally, propositional (or \"flat\") data representations have dominated the statistical sciences. These representations assume that data consist of independent and identically distributed (iid) entities which can be represented by a single data table. More recently, data scientists have increasingly focused on \"relational\" data sets that consist of interrelated, heterogeneous entities. However, relational learning and causal discovery are rarely combined. Relational representations are wholly absent from the literature where causality is discussed explicitly. Instead, the literature on causality that uses the framework of graphical models assumes that data are independent and identically distributed. This unexplored topical intersection represents an opportunity for advancement \u2014 by combining relational learning with causal reasoning, we can provide insight into the challenges found in each subject area. By adopting a causal viewpoint, we can clarify the mechanisms that produce previously identified pathologies in relational learning. Analogously, we can utilize relational data to establish and strengthen causal claims in ways that are impossible using only propositional representations.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('773','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2011\">2011<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Marc Maier, Matthew Rattigan, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('754','tp_abstract')\" style=\"cursor:pointer;\">Indexing Network Structure with Shortest-Path Trees<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">ACM Trans. Knowl. Discov. Data, <\/span><span class=\"tp_pub_additional_volume\">vol. 5, <\/span><span class=\"tp_pub_additional_number\">no. 3, <\/span><span class=\"tp_pub_additional_year\">2011<\/span>, <span class=\"tp_pub_additional_issn\">ISSN: 1556-4681<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_754\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('754','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_754\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('754','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_754\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('754','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_754\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{10.1145\/1993077.1993079,<br \/>\r\ntitle = {Indexing Network Structure with Shortest-Path Trees},<br \/>\r\nauthor = {Marc Maier and Matthew Rattigan and David Jensen},<br \/>\r\nurl = {https:\/\/doi.org\/10.1145\/1993077.1993079},<br \/>\r\ndoi = {10.1145\/1993077.1993079},<br \/>\r\nissn = {1556-4681},<br \/>\r\nyear  = {2011},<br \/>\r\ndate = {2011-08-01},<br \/>\r\njournal = {ACM Trans. Knowl. Discov. Data},<br \/>\r\nvolume = {5},<br \/>\r\nnumber = {3},<br \/>\r\npublisher = {Association for Computing Machinery},<br \/>\r\naddress = {New York, NY, USA},<br \/>\r\nabstract = {The ability to discover low-cost paths in networks has practical consequences for knowledge discovery and social network analysis tasks. Many analytic techniques for networks require finding low-cost paths, but exact methods for search become prohibitive for large networks, and data sets are steadily increasing in size. Short paths can be found efficiently by utilizing an index of network structure, which estimates network distances and enables rapid discovery of short paths. Through experiments on synthetic networks, we demonstrate that one such novel network structure index based on the shortest-path tree outperforms other previously proposed indices. We also show that it generalizes across arbitrarily weighted networks of various structures and densities, provides accurate estimates of distance, and has efficient time and space complexity. We present results on real data sets for several applications, including navigation, diameter estimation, centrality computation, and clustering---all made efficient by virtue of the network structure index.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('754','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_754\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The ability to discover low-cost paths in networks has practical consequences for knowledge discovery and social network analysis tasks. Many analytic techniques for networks require finding low-cost paths, but exact methods for search become prohibitive for large networks, and data sets are steadily increasing in size. Short paths can be found efficiently by utilizing an index of network structure, which estimates network distances and enables rapid discovery of short paths. Through experiments on synthetic networks, we demonstrate that one such novel network structure index based on the shortest-path tree outperforms other previously proposed indices. We also show that it generalizes across arbitrarily weighted networks of various structures and densities, provides accurate estimates of distance, and has efficient time and space complexity. We present results on real data sets for several applications, including navigation, diameter estimation, centrality computation, and clustering---all made efficient by virtue of the network structure index.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('754','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_754\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.1145\/1993077.1993079\" title=\"https:\/\/doi.org\/10.1145\/1993077.1993079\" target=\"_blank\">https:\/\/doi.org\/10.1145\/1993077.1993079<\/a><\/li><li><i class=\"ai ai-doi\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/dx.doi.org\/10.1145\/1993077.1993079\" title=\"Follow DOI:10.1145\/1993077.1993079\" target=\"_blank\">doi:10.1145\/1993077.1993079<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('754','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Matthew Rattigan, Marc Maier, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('752','tp_abstract')\" style=\"cursor:pointer;\">Relational blocking for causal discovery<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, <\/span><span class=\"tp_pub_additional_year\">2011<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_752\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('752','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_752\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('752','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_752\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('752','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_752\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{rattigan2011relational,<br \/>\r\ntitle = {Relational blocking for causal discovery},<br \/>\r\nauthor = {Matthew Rattigan and Marc Maier and David Jensen},<br \/>\r\nurl = {http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI11\/paper\/view\/3760},<br \/>\r\nyear  = {2011},<br \/>\r\ndate = {2011-01-01},<br \/>\r\nbooktitle = {Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence},<br \/>\r\nvolume = {25},<br \/>\r\nnumber = {1},<br \/>\r\nabstract = {Blocking is a technique commonly used in manual statistical analysis to account for confounding variables. However, blocking is not currently used in automated learning algorithms. These algorithms rely solely on statistical conditioning as an operator to identify conditional independence. In this work, we present relational blocking as a new operator that can be used for learning the structure of causal models. We describe how blocking is enabled by relational data sets, where blocks are determined by the links in the network. By blocking on entities rather than conditioning on variables, relational blocking can account for both measured and unobserved variables. We explain the mechanism of these methods using graphical models and the semantics of d-separation. Finally, we demonstrate the effectiveness of relational blocking for use in causal discovery by showing how blocking can be used in the causal analysis of two real-world social media systems.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('752','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_752\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Blocking is a technique commonly used in manual statistical analysis to account for confounding variables. However, blocking is not currently used in automated learning algorithms. These algorithms rely solely on statistical conditioning as an operator to identify conditional independence. In this work, we present relational blocking as a new operator that can be used for learning the structure of causal models. We describe how blocking is enabled by relational data sets, where blocks are determined by the links in the network. By blocking on entities rather than conditioning on variables, relational blocking can account for both measured and unobserved variables. We explain the mechanism of these methods using graphical models and the semantics of d-separation. Finally, we demonstrate the effectiveness of relational blocking for use in causal discovery by showing how blocking can be used in the causal analysis of two real-world social media systems.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('752','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_752\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI11\/paper\/view\/3760\" title=\"http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI11\/paper\/view\/3760\" target=\"_blank\">http:\/\/www.aaai.org\/ocs\/index.php\/AAAI\/AAAI11\/paper\/view\/3760<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('752','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Phillip B Kirlin, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('753','tp_abstract')\" style=\"cursor:pointer;\">Probabilistic Modeling of Hierarchical Music Analysis.<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Proceedings of the 12th International Society for Music Information \r\n Retrieval Conference, ISMIR, <\/span><span class=\"tp_pub_additional_pages\">pp. 393\u2013398, <\/span><span class=\"tp_pub_additional_year\">2011<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_753\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('753','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_753\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('753','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_753\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('753','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_753\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{kirlin2011probabilistic,<br \/>\r\ntitle = {Probabilistic Modeling of Hierarchical Music Analysis.},<br \/>\r\nauthor = {Phillip B Kirlin and David Jensen},<br \/>\r\nurl = {http:\/\/ismir2011.ismir.net\/papers\/PS3-7.pdf},<br \/>\r\nyear  = {2011},<br \/>\r\ndate = {2011-01-01},<br \/>\r\nbooktitle = {Proceedings of the 12th International Society for Music Information <br \/>\r\n Retrieval Conference, ISMIR},<br \/>\r\npages = {393--398},<br \/>\r\nabstract = {Hierarchical music analysis, as exemplified by Schenkerian analysis, describes the structure of a musical composition by a hierarchy among its notes. Each analysis defines a set of prolongations, where musical objects persist in time even though others are present. We present a formal model for representing hierarchical music analysis, probabilistic interpretations of that model, and an efficient algorithm for computing the most probable analysis under these interpretations. We represent Schenkerian analyses as maximal outerplanar graphs (MOPs). We use this representation to encode the largest known data set of computer-processable Schenkerian analyses, and we use these data to identify statistical regularities in the human-generated analyses. We show that a dynamic programming algorithm can be applied to these regularities to identify the maximum likelihood analysis for a given piece of music.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('753','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_753\" style=\"display:none;\"><div class=\"tp_abstract_entry\">Hierarchical music analysis, as exemplified by Schenkerian analysis, describes the structure of a musical composition by a hierarchy among its notes. Each analysis defines a set of prolongations, where musical objects persist in time even though others are present. We present a formal model for representing hierarchical music analysis, probabilistic interpretations of that model, and an efficient algorithm for computing the most probable analysis under these interpretations. We represent Schenkerian analyses as maximal outerplanar graphs (MOPs). We use this representation to encode the largest known data set of computer-processable Schenkerian analyses, and we use these data to identify statistical regularities in the human-generated analyses. We show that a dynamic programming algorithm can be applied to these regularities to identify the maximum likelihood analysis for a given piece of music.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('753','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_753\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-file-pdf\"><\/i><a class=\"tp_pub_list\" href=\"http:\/\/ismir2011.ismir.net\/papers\/PS3-7.pdf\" title=\"http:\/\/ismir2011.ismir.net\/papers\/PS3-7.pdf\" target=\"_blank\">http:\/\/ismir2011.ismir.net\/papers\/PS3-7.pdf<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('753','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_inproceedings\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Huseyin Oktay, A Soner Balkir, Ian Foster, David Jensen<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('756','tp_abstract')\" style=\"cursor:pointer;\">Distance estimation for very large networks using mapreduce and network structure indices<\/a> <span class=\"tp_pub_type tp_  inproceedings\">Proceedings Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_booktitle\">Workshop on Information Networks, <\/span><span class=\"tp_pub_additional_year\">2011<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_756\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('756','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_756\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('756','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_756\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@inproceedings{oktay2011distance,<br \/>\r\ntitle = {Distance estimation for very large networks using mapreduce and network structure indices},<br \/>\r\nauthor = {Huseyin Oktay and A Soner Balkir and Ian Foster and David Jensen},<br \/>\r\nyear  = {2011},<br \/>\r\ndate = {2011-01-01},<br \/>\r\nbooktitle = {Workshop on Information Networks},<br \/>\r\nabstract = {The ability to discover low-cost paths in networks has practical consequences for knowledge discovery and social network analysis tasks. Many analytic techniques for networks require finding low-cost paths, but exact methods for search become prohibitive for large networks, and data sets are steadily increasing in size. Short paths can be found efficiently by utilizing an index of network structure, which estimates network distances and enables rapid discovery of short paths. Through experiments on synthetic networks, we demonstrate that one such novel network structure index based on the shortest-path tree outperforms other previously proposed indices. We also show that it generalizes across arbitrarily weighted networks of various structures and densities, provides accurate estimates of distance, and has efficient time and space complexity. We present results on real data sets for several applications, including navigation, diameter estimation, centrality computation, and clustering---all made efficient by virtue of the network structure index.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {inproceedings}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('756','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_756\" style=\"display:none;\"><div class=\"tp_abstract_entry\">The ability to discover low-cost paths in networks has practical consequences for knowledge discovery and social network analysis tasks. Many analytic techniques for networks require finding low-cost paths, but exact methods for search become prohibitive for large networks, and data sets are steadily increasing in size. Short paths can be found efficiently by utilizing an index of network structure, which estimates network distances and enables rapid discovery of short paths. Through experiments on synthetic networks, we demonstrate that one such novel network structure index based on the shortest-path tree outperforms other previously proposed indices. We also show that it generalizes across arbitrarily weighted networks of various structures and densities, provides accurate estimates of distance, and has efficient time and space complexity. We present results on real data sets for several applications, including navigation, diameter estimation, centrality computation, and clustering---all made efficient by virtue of the network structure index.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('756','tp_abstract')\">Close<\/a><\/p><\/div><\/div><\/div><h3 class=\"tp_h3\" id=\"tp_h3_2010\">2010<\/h3><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Michael Hay, Gerome Miklau, David Jensen<\/p><p class=\"tp_pub_title\">Analyzing private network data <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">Privacy-aware knowledge discovery: Novel applications and new techniques, <\/span><span class=\"tp_pub_additional_pages\">pp. 459\u2013498, <\/span><span class=\"tp_pub_additional_year\">2010<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_768\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('768','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_768\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{hay2010analyzing,<br \/>\r\ntitle = {Analyzing private network data},<br \/>\r\nauthor = {Michael Hay and Gerome Miklau and David Jensen},<br \/>\r\nyear  = {2010},<br \/>\r\ndate = {2010-01-01},<br \/>\r\njournal = {Privacy-aware knowledge discovery: Novel applications and new techniques},<br \/>\r\npages = {459--498},<br \/>\r\npublisher = {Chapman & Hall\/CRC},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('768','tp_bibtex')\">Close<\/a><\/p><\/div><\/div><\/div><div class=\"tp_publication tp_publication_article\"><div class=\"tp_pub_info\"><p class=\"tp_pub_author\">Michael Hay, Gerome Miklau, David Jensen, Don Towsley, Chao Li<\/p><p class=\"tp_pub_title\"><a class=\"tp_title_link\" onclick=\"teachpress_pub_showhide('744','tp_abstract')\" style=\"cursor:pointer;\">Resisting structural re-identification in anonymized social networks<\/a> <span class=\"tp_pub_type tp_  article\">Journal Article<\/span> <\/p><p class=\"tp_pub_additional\"><span class=\"tp_pub_additional_in\">In: <\/span><span class=\"tp_pub_additional_journal\">The VLDB Journal, <\/span><span class=\"tp_pub_additional_volume\">vol. 19, <\/span><span class=\"tp_pub_additional_number\">no. 6, <\/span><span class=\"tp_pub_additional_pages\">pp. 797\u2013823, <\/span><span class=\"tp_pub_additional_year\">2010<\/span>.<\/p><p class=\"tp_pub_menu\"><span class=\"tp_abstract_link\"><a id=\"tp_abstract_sh_744\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('744','tp_abstract')\" title=\"Show abstract\" style=\"cursor:pointer;\">Abstract<\/a><\/span> | <span class=\"tp_resource_link\"><a id=\"tp_links_sh_744\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('744','tp_links')\" title=\"Show links and resources\" style=\"cursor:pointer;\">Links<\/a><\/span> | <span class=\"tp_bibtex_link\"><a id=\"tp_bibtex_sh_744\" class=\"tp_show\" onclick=\"teachpress_pub_showhide('744','tp_bibtex')\" title=\"Show BibTeX entry\" style=\"cursor:pointer;\">BibTeX<\/a><\/span><\/p><div class=\"tp_bibtex\" id=\"tp_bibtex_744\" style=\"display:none;\"><div class=\"tp_bibtex_entry\"><pre>@article{hay2010resisting,<br \/>\r\ntitle = {Resisting structural re-identification in anonymized social networks},<br \/>\r\nauthor = {Michael Hay and Gerome Miklau and David Jensen and Don Towsley and Chao Li},<br \/>\r\nurl = {https:\/\/doi.org\/10.1007\/s00778-010-0210-x},<br \/>\r\nyear  = {2010},<br \/>\r\ndate = {2010-01-01},<br \/>\r\njournal = {The VLDB Journal},<br \/>\r\nvolume = {19},<br \/>\r\nnumber = {6},<br \/>\r\npages = {797--823},<br \/>\r\npublisher = {Springer-Verlag},<br \/>\r\nabstract = {We identify privacy risks associated with releasing network data sets and provide an algorithm that mitigates those risks. A network consists of entities connected by links representing relations such as friendship, communication, or shared activity. Maintaining privacy when publishing networked data is uniquely challenging because an individual's network context can be used to identify them even if other identifying information is removed. In this paper, we quantify the privacy risks associated with three classes of attacks on the privacy of individuals in networks, based on the knowledge used by the adversary. We show that the risks of these attacks vary greatly based on network structure and size. We propose a novel approach to anonymizing network data that models aggregate network structure and then allows samples to be drawn from that model. The approach guarantees anonymity for network entities while preserving the ability to estimate a wide variety of network measures with relatively little bias.},<br \/>\r\nkeywords = {},<br \/>\r\npubstate = {published},<br \/>\r\ntppubtype = {article}<br \/>\r\n}<br \/>\r\n<\/pre><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('744','tp_bibtex')\">Close<\/a><\/p><\/div><div class=\"tp_abstract\" id=\"tp_abstract_744\" style=\"display:none;\"><div class=\"tp_abstract_entry\">We identify privacy risks associated with releasing network data sets and provide an algorithm that mitigates those risks. A network consists of entities connected by links representing relations such as friendship, communication, or shared activity. Maintaining privacy when publishing networked data is uniquely challenging because an individual's network context can be used to identify them even if other identifying information is removed. In this paper, we quantify the privacy risks associated with three classes of attacks on the privacy of individuals in networks, based on the knowledge used by the adversary. We show that the risks of these attacks vary greatly based on network structure and size. We propose a novel approach to anonymizing network data that models aggregate network structure and then allows samples to be drawn from that model. The approach guarantees anonymity for network entities while preserving the ability to estimate a wide variety of network measures with relatively little bias.<\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('744','tp_abstract')\">Close<\/a><\/p><\/div><div class=\"tp_links\" id=\"tp_links_744\" style=\"display:none;\"><div class=\"tp_links_entry\"><ul class=\"tp_pub_list\"><li><i class=\"fas fa-globe\"><\/i><a class=\"tp_pub_list\" href=\"https:\/\/doi.org\/10.1007\/s00778-010-0210-x\" title=\"https:\/\/doi.org\/10.1007\/s00778-010-0210-x\" target=\"_blank\">https:\/\/doi.org\/10.1007\/s00778-010-0210-x<\/a><\/li><\/ul><\/div><p class=\"tp_close_menu\"><a class=\"tp_close\" onclick=\"teachpress_pub_showhide('744','tp_links')\">Close<\/a><\/p><\/div><\/div><\/div><\/div><div class=\"tablenav\"><div class=\"tablenav-pages\"><span class=\"displaying-num\">127 entries<\/span> <a class=\"page-numbers button disabled\">&laquo;<\/a> <a class=\"page-numbers button disabled\">&lsaquo;<\/a> 1 of 3 <a href=\"https:\/\/groups.cs.umass.edu\/kdl\/publications\/?limit=2&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"next page\" class=\"page-numbers button\">&rsaquo;<\/a> <a href=\"https:\/\/groups.cs.umass.edu\/kdl\/publications\/?limit=3&amp;tgid=&amp;yr=&amp;type=&amp;usr=&amp;auth=&amp;tsr=#tppubs\" title=\"last page\" class=\"page-numbers button\">&raquo;<\/a> <\/div><\/div><\/div>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":47,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-189","page","type-page","status-publish","hentry","hfeed"],"_links":{"self":[{"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/pages\/189","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/users\/47"}],"replies":[{"embeddable":true,"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/comments?post=189"}],"version-history":[{"count":43,"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/pages\/189\/revisions"}],"predecessor-version":[{"id":261,"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/pages\/189\/revisions\/261"}],"wp:attachment":[{"href":"https:\/\/groups.cs.umass.edu\/kdl\/wp-json\/wp\/v2\/media?parent=189"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}