Search

Search Funnelback University

Search powered by Funnelback
1 - 15 of 15 search results for People aliens |u:mlg.eng.cam.ac.uk where 0 match all words and 15 match some words.
  1. Results that match 1 of 2 words

  2. 19 Jun 2024: on their effects on people belonging to certain sensitivedemographic groups (e.g., gender, race). ... Now we compare user judgments of fairness of individual features, fordifferent demographic subgroups of people.
  3. Human Perceptions of Fairness in Algorithmic Decision Making: A Case…

    https://mlg.eng.cam.ac.uk/adrian/WWW18-HumanPerceptions.pdf
    19 Jun 2024: Our Contributions. We collected and analyzed fairness judgmentsfrom a survey of 576 people. ... So our framework for how people judgefeature usage fairness consists of two-parts.
  4. Transparency: Motivations and Challenges? Adrian…

    https://mlg.eng.cam.ac.uk/adrian/transparency.pdf
    19 Jun 2024: Yet both concepts are somewhat ambiguous, and can meandifferent things to different people in different contexts. ... People might be experts or not. We list several types and goals of transparency.
  5. Beyond Distributive Fairness in Algorithmic Decision Making: Feature…

    https://mlg.eng.cam.ac.uk/adrian/AAAI18-BeyondDistributiveFairness.pdf
    19 Jun 2024: However, we note that such judgments can begathered from any other group of people, ranging fromcrowd workers to domain experts. ... people) more likelyto be falsely predicted as having a higher risk of recidivismthan another group of people (e.g., white
  6. You Shouldn’t Trust Me: Learning Models WhichConceal Unfairness From…

    https://mlg.eng.cam.ac.uk/adrian/ECAI20-You_Shouldn%E2%80%99t_Trust_Me.pdf
    19 Jun 2024: Let x2 be a sensitive fea-ture, such as age, given by the shape of the point: assume youngand mature people. ... 100% points are accurate (correctly, blue maturepeople are in the blue zone, red mature people are in the red zone).
  7. Network Ranking With Bethe Pseudomarginals Kui TangColumbia…

    https://mlg.eng.cam.ac.uk/adrian/2013_NeurIPS_DiscML_Network.pdf
    19 Jun 2024: 1 Introduction. Many important data-sets involve networks: people belong to social networks, webpages are joined in a linkgraph, and power utilities are connected in a grid.
  8. Adversarial Graph Embeddings for Fair Influence Maximization over…

    https://mlg.eng.cam.ac.uk/adrian/IJCAI20_AdversarialGraphEmbeddings.pdf
    19 Jun 2024: in terms of thetotal number of people influenced, and the fraction of peopleinfluenced from both A and B communities. ... maximizing the to-tal number of influenced people, while also significantly de-creasing disparity.
  9. 2018 Formatting Instructions for Authors Using LaTeX

    https://mlg.eng.cam.ac.uk/adrian/AIES18-crowd_signals.pdf
    19 Jun 2024: 2017),and might help to promote greater understanding and cohe-sion among people. ... CNN You mean, like the UNFOUNDED claims of Russiancollusion? You people are typically selective in your bias pro?https://t.co/CESkVpIZOk@nytimes His actions were
  10. ML-IRL: Machine Learning in Real Life Workshop at ICLR ...

    https://mlg.eng.cam.ac.uk/adrian/ML_IRL_2020-CLUE.pdf
    19 Jun 2024: SampleSize” denotes how many people received this variant. The next two columns contain the proportionof correct answers when identifying epistemic or aleatoric uncertainty for LSAT, respectively.
  11. Bounding the Integrality Distance ofLP Relaxations for Structured…

    https://mlg.eng.cam.ac.uk/adrian/OPT2016_paper_3.pdf
    19 Jun 2024: That is, max-margin training with approximateinference—which is something people do anyway to learn graphical models—reduces not only theprediction error, but also the inference approximation error.
  12. Blind Justice: Fairness with Encrypted Sensitive Attributes

    https://mlg.eng.cam.ac.uk/adrian/ICML18-BlindJustice.pdf
    19 Jun 2024: 2.2. Fairness Criteria. In large part, works that formalize fairness in machine learn-ing do so by balancing a certain condition between groupsof people with different sensitive attributes, z versus ... Figure 3. The fraction of people with z = 0
  13. One-network Adversarial Fairness

    https://mlg.eng.cam.ac.uk/adrian/AAAI2019_OneNetworkAdversarialFairness.pdf
    19 Jun 2024: Here, following earlierwork, unfairness means discriminating against a particulargroup of people due to sensitive group characteristics suchas gender or race (Grgic-Hlaca et al. ... For the Adult data, a random guessing classi-fier would result in 75.9%
  14. A Unified Approach to Quantifying Algorithmic Unfairness: Measuring…

    https://mlg.eng.cam.ac.uk/adrian/KDD2018_inequality_indices.pdf
    19 Jun 2024: We furtherillustrate this point with two examples: First, by this definition amodel that assigns the same outcome to everyone is considered fair,regardless of people’s merit for different outcomes ... This offersa framework to interpolate between group
  15. Working Draft 1 Accountability of AI Under the Law: ...

    https://mlg.eng.cam.ac.uk/adrian/SSRN-id3064761-Dec19.pdf
    19 Jun 2024: for most people.8 Moreover, decisions about how to define objective functions and what training data to use can introduce human error into AI decision making.9 Thus, there exist legitimate
  16. Methods for Inference in Graphical Models

    https://mlg.eng.cam.ac.uk/adrian/phd_FINAL.pdf
    19 Jun 2024: I’ve had the privilege to collaborate with wonderful people. Many thanks to Tony, Dan Ellis,.

Refine your results

Search history

Recently clicked results

Recently clicked results

Your click history is empty.

Recent searches

Recent searches

Your search history is empty.