Home About Subscribe Search Member Area

Humanist Discussion Group

< Back to Volume 32

Humanist Archives: Feb. 11, 2019, 5:53 a.m. Humanist 32.445 - events: fairness in user modelling

                  Humanist Discussion Group, Vol. 32, No. 445.
            Department of Digital Humanities, King's College London
                   Hosted by King's Digital Lab
                Submit to: humanist@dhhumanist.org

        Date: 2019-02-10 17:31:32+00:00
        From: Robin Burke 
        Subject: Call for papers: Second Workshop on Fairness in User Modeling, Adaptation, and Personalization (FairUMAP 2019)

Fairness in user modeling may be of interest to some in this group.


Second Workshop on Fairness in User Modeling, Adaptation, and Personalization
(FairUMAP 2019)

At the ACM Conference on User Modeling, Adaptation, and Personalization (UMAP
9 -- 12 June, 2019, Larnaca, Cyprus

Workshop website: https://fairumap.wordpress.com/
Conference website: http://www.cyprusconferences.org/umap2019/


Personalization has become a ubiquitous and essential part of systems that help
users find relevant information in today's highly complex information-rich
online environments. Machine learning, recommender systems, and user modeling
are key enabling technologies that allow intelligent systems to learn from users
and adapt their output to users' needs and preferences. However, there has been
a growing recognition that these underlying technologies raise novel ethical,
legal, and policy challenges. It has become apparent that a single-minded focus
on the user preferences has obscured other important and beneficial outcomes
such systems must be able to deliver. System properties such as fairness,
transparency, balance, openness to diversity, and other social welfare
considerations are not captured by typical metrics based on which data-driven
personalized models are optimized. Indeed, widely-used personalization systems
in such popular sites such as Facebook, Google News and YouTube have been
heavily criticized for personalizing information delivery too heavily at the
cost of these other objectives.

Bias, fairness, and transparency in machine learning are topics of considerable
recent research interest. However, more work is needed to expand and extend this
work into algorithmic and modeling approaches where personalization and user
modeling are of primary importance. In particular, it is essential to address
these challenges from the standpoint of understanding stereotypes in users’
behaviors and their influence on user or group decisions.

The 2nd Workshop on Fairness in User Modeling, Adaptation, and Personalization
aims to bring together experts from academia and industry to discuss ethical,
social, and legal concerns related to personalization and user modeling with the
goal of exploring a variety of mechanisms and modeling approaches that help
mitigate bias and achieve fairness in personalized systems.


Topics of interest include, but are not limited to the following.

- Bias and discrimination in user modeling, personalization and recommendation
- Computational techniques and algorithms for fairness-aware personalization
- Definitions, metrics and criteria for optimizing and evaluating fairness-
related aspects of personalized systems
- Data preprocessing and transformation methods to address bias in training data
- User modeling approaches that take fairness and bias into account
- User studies and other empirical studies to evaluate impact of personalization
on fairness, balance, diversity, and other social welfare criteria
- Balancing needs of multiple stakeholders in recommender systems and other
personalized systems
- "Filter bubble" or "balkanization" effects of personalization
- Transparent and accurate explanations for recommendations and other
personalization outcomes


Submission deadline: March 13, 2019 (23:59 American Samoa Zone - UTC-11)
Notification of acceptance: March 26, 2019
Camera-ready due: April 3, 2019 (23:59 American Samoa Zone - UTC-11)


Research papers reporting original results as well as position papers proposing
novel and ground-breaking ideas pertaining to the workshop topics are solicited.

Manuscripts must be in English with a maximum length of 6 pages for research
papers and 3 pages for position papers (with a maximum of one additional page
for references). Papers must be formatted using the ACM SIG Standard (SIGCONF)
proceedings template: https://www.acm.org/publications/proceedings-template.
Accepted papers will be either presented as a talk or poster (to be determined).

At least one author of each accepted paper must attend the workshop and present
the paper.

Please submit your paper by using the on-line submission system via:

WORKSHOP Organizing Committee

Bamshad Mobasher, DePaul University, USA
Robin Burke, University of Colorado, Boulder, USA
Michael Ekstrand, Boise State University, USA
Bettina Berendt, KU Leuven, Belgium
Styliani Kleanthous, Open University of Cyprus
Tsvi Kuflik, University of Haifa, Israel
Jahna Otterbacher, Open University of Cyprus
Alan Hartman, University of Haifa, Israel
Veronika Bogina, University of Haifa, Israel

Robin Burke, Professor
Department of Information Science
Department of Computer Science (by courtesy)
University of Colorado, Boulder

Unsubscribe at: http://dhhumanist.org/Restricted
List posts to: humanist@dhhumanist.org
List info and archives at at: http://dhhumanist.org
Listmember interface at: http://dhhumanist.org/Restricted/
Subscribe at: http://dhhumanist.org/membership_form.php

Editor: Willard McCarty (King's College London, U.K.; Western Sydney University, Australia)
Software designer: Malgosia Askanas (Mind-Crafts)

This site is maintained under a service level agreement by King's Digital Lab.