Sandip soparkar wiki biography of rory
•
Yujia Li(李宇佳)
2019CVPR outstanding reviewer
2016ICLR travel award
2015Microsoft Ph.D. fellowship (US and Canada) finalist
2015ICML travel grant
2013CVPR travel grant
2013University of Toronto School of Graduate Studies conference grant
2008-2010University-wide comprehensive merit scholarship, three times - including Kai-Feng Scholarship, which is the highest amount among all scholarships and awarded to only 30 undergraduate students in Tsinghua University every year.
20082nd Prize - kinesisk National College Physics Contest
20061st Prize - Chinese Physics Olympiad (CPhO) in Provinces
20052nd Prize - Chinese National Olympiad in Informatics in Provinces (NOIP)
Papers共 54 篇Author StatisticsCo-AuthorSimilar Experts
•
Man Made Language Models? Evaluating LLMs’ Perpetuation of Masculine Generics Bias
Enzo Doyen
University of Strasbourg
enzo.doyen@unistra.fr &Amalia Todirascu
University of Strasbourg
todiras@unistra.fr
Abstract
Large language models (LLMs) have been shown to propagate and even amplify gender bias, in English and other languages, in specific or constrained contexts. However, no studies so far have focused on gender biases conveyed by LLMs’ responses to generic instructions, especially with regard to masculine generics (MG). MG are a linguistic feature found in many gender-marked languages, denoting the use of the masculine gender as a “default” or supposedly neutral gender to refer to mixed group of men and women, or of a person whose gender is irrelevant or unknown. Numerous psycholinguistics studies have shown that MG are not neutral and induce gender bias. This work aims to analyze the use of MG by both proprietary and local LLMs in responses to generic instructions a
•
Towards Geo-Culturally Grounded LLM Generations
Piyawat Lertvittayakumjorn†, David Kinney⋆†‡,
Vinodkumar Prabhakaran†, Donald Martin†, Sunipa Dev†
†Google ‡Washington University in St. Louis
{piyawat,vinodkpg,dxm,sunipadev}@google.com, kinney@wustl.edu
Abstract
Generative large language models (LLMs) have been demonstrated to have gaps in diverse, cultural knowledge across the globe. We investigate the effect of retrieval augmented generation and search-grounding techniques on the ability of LLMs to display familiarity with a diverse range of national cultures. Specifically, we compare the performance of standard LLMs, LLMs augmented with retrievals from a bespoke knowledge base (i.e., KB grounding), and LLMs augmented with retrievals from a web search (i.e., search grounding) on a series of cultural familiarity benchmarks. We find that search grounding significantly improves the LLM performance on multiple-choice benchmarks that test propositional knowledge (