All Stories

  1. Ideology-Based LLMs for Content Moderation
  2. Scalable Methods for Storing and Retrieving Wikipedia Revision Histories for Large-Scale Analysis
  3. Natural Language Processing for the Legal Domain: A Survey of Tasks, Datasets, Models, and Challenges
  4. Exploring Wikipedia Gender Diversity Over Time — The Wikipedia Gender Dashboard (WGD)
  5. Preaching to the ChoIR: Lessons IR Should Share with AI
  6. Bias in Humans and AI - What To Do About It?
  7. BiasNavi: LLM-Empowered Data Bias Management
  8. The Impact of Persona-based Political Perspectives on Hateful Content Detection
  9. Mapping and Influencing the Political Ideology of Large Language Models using Synthetic Personas
  10. Crowdsourcing or AI Sourcing?
  11. Optimizing LLMs with Direct Preferences: A Data Efficiency Perspective
  12. Hate Speech Detection with Generalizable Target-aware Fairness
  13. Understanding the Barriers to Running Longitudinal Studies on Crowdsourcing Platforms
  14. Fairness without Sensitive Attributes via Knowledge Sharing
  15. How Good are LLMs in Generating Personalized Advertisements?
  16. Who Determines What Is Relevant? Humans or AI? Why Not Both?
  17. Editorial: Special Issue on Human in the Loop Data Curation
  18. On the Impact of Showing Evidence from Peers in Crowdsourced Truthfulness Assessments
  19. Data Bias Management
  20. Perspectives on Large Language Models for Relevance Judgment
  21. On the Impact of Data Quality on Image Classification Fairness
  22. How Many Crowd Workers Do I Need? On Statistical Power When Crowdsourcing Relevance Judgments
  23. Human-in-the-loop Regular Expression Extraction for Single Column Format Inconsistency
  24. The Community Notes Observatory: Can Crowdsourced Fact-Checking be Trusted in Practice?
  25. Report on the 1st Workshop on Human-in-the-Loop Data Curation (HIL-DC 2022) at CIKM 2022
  26. A Data-Driven Analysis of Behaviors in Data Curation Processes
  27. Combining Human and Machine Confidence in Truthfulness Assessment
  28. Using Computers to Fact-Check Text and Justify the Decision
  29. Socio-Economic Diversity in Human Annotations
  30. Preferences on a Budget: Prioritizing Document Pairs when Crowdsourcing Relevance Judgments
  31. Does Evidence from Peers Help Crowd Workers in Assessing Truthfulness?
  32. Effects of Technological Interventions for Self-regulation: A Control Experiment in Learnersourcing
  33. Hierarchical Clustering of Corals using Image Clustering
  34. An Analysis of the Australian Political Discourse in Sponsored Social Media Content
  35. On the State of Reporting in Crowdsourcing Experiments and a Checklist to Aid Current Practices
  36. Charting the Design and Analytics Agenda of Learnersourcing Systems
  37. Report on the first workshop on bias in automatic knowledge graph construction at AKBC 2020
  38. Modelling User Behavior Dynamics with Embeddings
  39. The COVID-19 Infodemic
  40. How to make crowd workers earn an hourly wage
  41. On Understanding Data Worker Interaction Behaviors
  42. Can The Crowd Identify Misinformation Objectively?
  43. Representation learning for entity type ranking
  44. Health Card Retrieval for Consumer Health Search
  45. On Transforming Relevance Scales
  46. Understanding Worker Moods and Reactions to Rejection in Crowdsourcing
  47. Quality Control Attack Schemes in Crowdsourcing
  48. Health Cards for Consumer Health Search
  49. Implicit Bias in Crowdsourced Knowledge Graphs
  50. Scalpel-CD: Leveraging Crowdsourcing and Deep Probabilistic Modeling for Debugging Noisy Training Data
  51. Deadline-Aware Fair Scheduling for Multi-Tenant Crowd-Powered Systems
  52. All Those Wasted Hours
  53. Novel insights into views towards H1N1 during the 2009 Pandemic: a thematic analysis of Twitter data
  54. Non-parametric Class Completeness Estimators for Collaborative Knowledge Graphs—The Case of Wikidata
  55. Semantic Interlinking
  56. The Impact of Task Abandonment in Crowdsourcing
  57. The Evolution of Power and Standard Wikidata Editors: Comparing Editing Behavior over Time to Predict Lifespan and Volume of Edits
  58. Can User Behaviour Sequences Reflect Perceived Novelty?
  59. Moral Panic through the Lens of Twitter
  60. Investigating User Perception of Gender Bias in Image Search
  61. On Fine-Grained Relevance Scales
  62. On the Volatility of Commercial Search Engines and its Impact on Information Retrieval Research
  63. Crowd Anatomy Beyond the Good and Bad: Behavioral Traces for Crowd Worker Modeling and Pre-selection
  64. Measuring the Effect of Public Health Campaigns on Twitter: The Case of World Autism Awareness Day
  65. Augmenting Intelligence with Humans-in-the-Loop (HumL@WWW2018) Chairs' Welcome & Organization
  66. Chapter 4: Using Twitter as a Data Source: An Overview of Ethical, Legal, and Methodological Challenges
  67. Understanding Engagement through Search Behaviour
  68. Considering Assessor Agreement in IR Evaluation
  69. Modus Operandi of Crowd Workers
  70. An Introduction to Hybrid Human-Machine Information Systems
  71. Towards building a standard dataset for Arabic keyphrase extraction evaluation
  72. Scheduling Human Intelligence Tasks in Multi-Tenant Crowd-Powered Systems
  73. Contextualized ranking of entity types based on knowledge graphs
  74. A Tutorial on Leveraging Knowledge Graphs for Web Search
  75. The Relationship Between User Perception and User Behaviour in Interactive Information Retrieval Evaluation
  76. Hybrid human–machine information systems: Challenges and opportunities
  77. Pooling-based continuous evaluation of information retrieval systems
  78. Human Beyond the Machine: Challenges and Opportunities of Microtask Crowdsourcing
  79. The Dynamics of Micro-Task Crowdsourcing
  80. Understanding Malicious Behavior in Crowdsourcing Platforms
  81. Correct Me If I'm Wrong
  82. B-hist: Entity-centric search over personal web browsing history
  83. Hippocampus
  84. Effective named entity recognition for idiosyncratic web collections
  85. Entity disambiguation in tweets leveraging user social profiles
  86. Large-scale linked data integration using probabilistic reasoning and crowdsourcing
  87. NoizCrowd: A Crowd-Based Data Gathering and Management System for Noise Level Data
  88. Ontology-Based Word Sense Disambiguation for Scientific Literature
  89. TRank: Ranking Entity Types Using the Web of Data
  90. The Bowlogna ontology: Fostering open curricula and agile knowledge bases for Europe's higher education landscape
  91. ZenCrowd
  92. BowlognaBench—Benchmarking RDF Analytics
  93. Combining inverted indices and structured search for ad-hoc object retrieval
  94. Predicting the Future Impact of News Events
  95. From people to entities
  96. Visual interfaces for stimulating exploratory search
  97. Report on INEX 2009
  98. Why finding entities in Wikipedia is difficult, sometimes
  99. Leveraging personal metadata for Desktop search: The Beagle++ system
  100. Dear search engine: what's your opinion about...?
  101. Entity summarization of news articles
  102. Exploiting click-through data for entity retrieval
  103. Overview of the INEX 2009 Entity Ranking Track
  104. Ranking Entities Using Web Search Query Logs
  105. TAER
  106. The missing links
  107. An Architecture for Finding Entities on the Web
  108. Report on INEX 2008
  109. A Vector Space Model for Ranking Entities and Its Application to Expert Search
  110. How to Trace and Revise Identities
  111. L3S at INEX 2008: Retrieving Entities Using Structured Information
  112. Overview of the INEX 2008 Entity Ranking Track
  113. A Model for Ranking Entities and Its Application to Wikipedia
  114. Social recommendations of content and metadata
  115. Leveraging semantic technologies for enterprise search
  116. A Classification of IR Effectiveness Metrics
  117. L3S at INEX 2007: Query Expansion for Entity Ranking Using a Highly Accurate Ontology
  118. Ranking Categories for Web Search
  119. Semantically Enhanced Entity Ranking