{"id":815,"date":"2023-11-22T16:46:55","date_gmt":"2023-11-22T16:46:55","guid":{"rendered":"https:\/\/www.wsdm-conference.org\/2024\/?page_id=815"},"modified":"2024-03-05T20:39:27","modified_gmt":"2024-03-05T20:39:27","slug":"tutorials","status":"publish","type":"page","link":"https:\/\/www.wsdm-conference.org\/2024\/tutorials\/","title":{"rendered":"Tutorials"},"content":{"rendered":"\n<div id=\"section-g00342b\" class=\"wp-block-gutentor-e1 section-g00342b gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\"><strong>Practical Bandits: An Industry Perspective<\/strong><\/h5><\/div><\/div>\n\n\n\n<div id=\"section-gce13ec\" class=\"wp-block-gutentor-e1 section-gce13ec gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Bram van den Akker (Booking), Olivier Jeunen (ShareChat), Ying Li (Netflix), Ben London (Amazon), Zahra Nazari (Spotify) and Devesh Parekh (Netflix)<\/p><\/div><\/div>\n\n\n\n<div id=\"section-gb7f50a\" class=\"wp-block-gutentor-e1 section-gb7f50a gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>The bandit paradigm provides a unified modeling framework for problems that require decision-making under uncertainty. Because many business metrics can be viewed as rewards (a.k.a. utilities) that result from actions, bandit algorithms have seen a large and growing interest from industrial applications, such as search, recommendation and advertising. Indeed, with the bandit lens comes the promise of direct optimization for the metrics we care about. Nevertheless, the road to successfully applying bandits in production is not an easy one. Even when the action space and rewards are well-defined, practitioners still need to make decisions regarding multi-arm or contextual approaches, on- or off-policy setups, delayed or immediate feedback, myopic or long-term optimisation, etc. To make matters worse, industrial platforms typically give rise to large action spaces in which existing approaches tend to break down. The research literature on these topics is broad and vast, but this can overwhelm practitioners, whose primary aim is to solve practical problems, and therefore need to decide on a specific instantiation or approach for each project. This tutorial will take a step towards filling that gap between the theory and practice of bandits. Our goal is to present a unified overview of the field and its existing terminology, concepts and algorithms&#8212;with a focus on problems relevant to industry. We hope our industrial perspective will help future practitioners who wish to leverage the bandit paradigm for their application.<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g146848\" class=\"wp-block-gutentor-e1 section-g146848 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\">Some Useful Things to Know When Combining IR and NLP: the Easy, the Hard and the Ugly<\/h5><\/div><\/div>\n\n\n\n<div id=\"section-g13731f\" class=\"wp-block-gutentor-e1 section-g13731f gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Omar Alonso (Amazon) and Kenneth Church (Northeastern University)<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g1cfb46\" class=\"wp-block-gutentor-e1 section-g1cfb46 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>Deep nets such as GPT are at the core of the current advances in many systems and applications. Things are moving fast; techniques become obsolete quickly (within weeks). How can we take advantage of new discoveries and incorporate them into our existing work? Are new developments radical improvements, or incremental repetitions of established concepts, or combinations of both? In this tutorial, we aim to bring interested researchers and practitioners up to speed on the recent and ongoing techniques around ML and Deep learning in the context of IR and NLP. Additionally, our goal is to clarify terminology, emphasize fundamentals, and outline problems and new research opportunities.<\/p><\/div><\/div>\n\n\n\n<div id=\"section-gde3b56\" class=\"wp-block-gutentor-e1 section-gde3b56 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\">Bridging Text Data and Graph Data: Towards Semantics and Structure-aware Knowledge Discovery<\/h5><\/div><\/div>\n\n\n\n<div id=\"section-g7c2bf8\" class=\"wp-block-gutentor-e1 section-g7c2bf8 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Bowen Jin, Yu Zhang, Sha Li and Jiawei Han (UIUC)<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g2f4076\" class=\"wp-block-gutentor-e1 section-g2f4076 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>Graphs and texts play crucial roles in data mining, each possessing unique characteristics that often require distinct modeling methods. Technologies for mining graph data and text data are usually designed separately. Nevertheless, frequently, data contains a blend of both modalities, with their information frequently complementing each other. For instance, in e-commerce data, the product-user graph and product descriptions provide distinct insights into product features. Similarly, in scientific literature, the citation graph, author information, and the content of papers collectively contribute to modeling the impact of a paper.<br>In this tutorial, our emphasis will be on exploring the latest advancements in graph mining techniques that leverage the capabilities of Pre-trained Language Models (PLMs), as well as the enhancement of text mining methods through the incorporation of graph structure information. We will present an organized picture of how graphs and texts can mutually benefit each other and lead to deeper knowledge discovery, with the following outline: (1) an introduction to how graph and text are intertwined in real-life data and how graph neural networks and pre-trained language models are designed to capture signal from graph and text modalities, (2) graph construction from text: construct sentence-level graphs, event graphs, reasoning, knowledge graphs from text, (3) network mining with language models: language model-based methods for representation learning on graph and language model pretraining on graphs, (4) text mining with structure information: text classification, literature retrieval, and question answering with graph structure as auxiliary information, (5) towards an integrated semantics and structure mining paradigm.<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g5aa044\" class=\"wp-block-gutentor-e1 section-g5aa044 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\">Unbiased Learning to Rank: On Recent Advances and Practical Applications<\/h5><\/div><\/div>\n\n\n\n<div id=\"section-g53a11c\" class=\"wp-block-gutentor-e1 section-g53a11c gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Shashank Gupta (UvA), Philipp Hager (UvA), Jin Huang (UvA), Ali Vardasbi (UvA) and Harrie Oosterhuis (Radboud University)<\/p><\/div><\/div>\n\n\n\n<div id=\"section-gd0799b\" class=\"wp-block-gutentor-e1 section-gd0799b gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>Since its inception, the field of unbiased learning to rank (ULTR) has remained very active and has seen several impactful advancements in recent years. This tutorial provides both an introduction to the core concepts of the field and an overview of recent advancements in its foundations, along with several applications of its methods. The tutorial is divided into four parts: Firstly, we give an overview of the different forms of bias that can be addressed with ULTR methods. Secondly, we present a comprehensive discussion of the latest estimation techniques in the ULTR field.Thirdly, we survey published results of ULTR in real-world applications. Fourthly, we discuss the connection between ULTR and fairness in ranking. We end by briefly reflecting on the future of ULTR research and its applications. This tutorial is intended to benefit both researchers and industry practitioners interested in developing new ULTR solutions or utilizing them in real-world applications.<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g20cd81\" class=\"wp-block-gutentor-e1 section-g20cd81 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\">Strategic ML: How to Learn With Data That \u2018Behaves\u2019<\/h5><\/div><\/div>\n\n\n\n<div id=\"section-g3fa1d8\" class=\"wp-block-gutentor-e1 section-g3fa1d8 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Nir Rosenfeld (Technion)<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g0a702a\" class=\"wp-block-gutentor-e1 section-g0a702a gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>The success of machine learning across a wide array of tasks and applications has made it appealing to use it also in the social domain. Indeed, learned models now form the backbone of recommendation systems, social media platforms, online markets, and e-commerce services, where they are routinely used to inform decisions by, for, and about their human users. But humans are not your conventional input&#8211;they have goals, beliefs, and aspirations, and take action to promote their own interests. Given that standard learning methods are not designed to handle inputs that `behave&#8217;, a natural question is: how should we design learning systems when we know they will be deployed and used in social settings? This tutorial introduces strategic machine learning, a new and emerging subfield of machine learning that aims to develop a disciplined framework for learning under strategic user behavior. The working hypothesis of strategic ML is simple: users want things, and act to achieve them. Surprisingly, this basic truism is difficult to address within the conventional learning framework. The key challenge is that how users behave often depends on the learned decision rule itself; thus, strategic learning seeks to devise methods which are able to anticipate and accommodate such responsive behavior. Towards this, strategic ML offers a formalism for reasoning about strategic responses, for designing appropriate learning objectives, and for developing practical tools for learning in strategic environments. The tutorial will survey recent and ongoing work in this new domain, present key theoretical and empirical results, provide practical tools, and discuss open questions and landmark challenges.<\/p><\/div><\/div>\n\n\n\n<div id=\"section-ga6f634\" class=\"wp-block-gutentor-e1 section-ga6f634 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\">Responsible AI<\/h5><\/div><\/div>\n\n\n\n<div id=\"section-gf8f76c\" class=\"wp-block-gutentor-e1 section-gf8f76c gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Ricardo Baeza Yates (Northeastern University) <\/p><\/div><\/div>\n\n\n\n<div id=\"section-g6427e7\" class=\"wp-block-gutentor-e1 section-g6427e7 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>In the first part of this tutorial we define responsible AI and we discuss the problems embedded in terms like ethical or trustworthy AI. In the second part, to set the stage, we cover irresponsible AI: discrimination (e.g., the impact of human biases); pseudo-science (e.g., biometric based behavioral predictions); human limitations (e.g., human incompetence, cognitive biases); technical limitations (data as a proxy of reality, wrong evaluation); social impact (e.g., unfair digital markets or mental health and disinformation issues created by large language models); environmental impact (e.g., indiscriminate use of computing resources). These examples do have a personal bias but set the context for the third part where we cover the current challenges: ethical principles, governance and regulation. We finish by discussing our responsible AI initiatives, many recommendations, and some philosophical issues.<\/p><\/div><\/div>\n\n\n\n<div id=\"section-gf7d395\" class=\"wp-block-gutentor-e1 section-gf7d395 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><h5 class=\"gutentor-text\">Trustworthy LLMs<\/h5><\/div><\/div>\n\n\n\n<div id=\"section-gbf1253\" class=\"wp-block-gutentor-e1 section-gbf1253 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Sanmi Koyejo (Stanford University) and Bo Li (U Chicago)<\/p><\/div><\/div>\n\n\n\n<div id=\"section-g1aedc2\" class=\"wp-block-gutentor-e1 section-g1aedc2 gutentor-element gutentor-element-advanced-text\"><div class=\"gutentor-text-wrap\"><p class=\"gutentor-text\">Abstract:<br>Large Language models are among the most exciting technologies developed in the last few years. While the model&#8217;s capabilities continue to improve, researchers, practitioners, and the general public are increasingly aware of some of its shortcomings. What will it take to build trustworthy large language models? This tutorial will present a range of recent findings, discussions, questions, and partial answers in the space of trustworthiness in large language models. While this tutorial will not attempt a comprehensive overview of this rich area, we aim to provide the participants with some tools and insights and to understand both the conceptual foundations of trustworthiness and a broad range of ongoing research efforts. We will tackle some of the hard questions that you may have about trustworthy large language models and hopefully address some misconceptions that have become pervasive.&nbsp;<br><\/p><\/div><\/div>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":1,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"footnotes":""},"class_list":["post-815","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/pages\/815","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/comments?post=815"}],"version-history":[{"count":11,"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/pages\/815\/revisions"}],"predecessor-version":[{"id":1206,"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/pages\/815\/revisions\/1206"}],"wp:attachment":[{"href":"https:\/\/www.wsdm-conference.org\/2024\/wp-json\/wp\/v2\/media?parent=815"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}