All Stories

  1. Scalable Methods for Storing and Retrieving Wikipedia Revision Histories for Large-Scale Analysis
  2. Natural Language Processing for the Legal Domain: A Survey of Tasks, Datasets, Models, and Challenges
  3. Exploring Wikipedia Gender Diversity Over Time — The Wikipedia Gender Dashboard (WGD)
  4. Preaching to the ChoIR: Lessons IR Should Share with AI
  5. Bias in Humans and AI - What To Do About It?
  6. BiasNavi: LLM-Empowered Data Bias Management
  7. The Impact of Persona-based Political Perspectives on Hateful Content Detection
  8. Mapping and Influencing the Political Ideology of Large Language Models using Synthetic Personas
  9. Crowdsourcing or AI Sourcing?
  10. Optimizing LLMs with Direct Preferences: A Data Efficiency Perspective
  11. Hate Speech Detection with Generalizable Target-aware Fairness
  12. Understanding the Barriers to Running Longitudinal Studies on Crowdsourcing Platforms
  13. Fairness without Sensitive Attributes via Knowledge Sharing
  14. How Good are LLMs in Generating Personalized Advertisements?
  15. Who Determines What Is Relevant? Humans or AI? Why Not Both?
  16. Editorial: Special Issue on Human in the Loop Data Curation
  17. On the Impact of Showing Evidence from Peers in Crowdsourced Truthfulness Assessments
  18. Data Bias Management
  19. Perspectives on Large Language Models for Relevance Judgment
  20. On the Impact of Data Quality on Image Classification Fairness
  21. How Many Crowd Workers Do I Need? On Statistical Power When Crowdsourcing Relevance Judgments
  22. Human-in-the-loop Regular Expression Extraction for Single Column Format Inconsistency
  23. The Community Notes Observatory: Can Crowdsourced Fact-Checking be Trusted in Practice?
  24. Report on the 1st Workshop on Human-in-the-Loop Data Curation (HIL-DC 2022) at CIKM 2022
  25. A Data-Driven Analysis of Behaviors in Data Curation Processes
  26. Combining Human and Machine Confidence in Truthfulness Assessment
  27. Using Computers to Fact-Check Text and Justify the Decision
  28. Socio-Economic Diversity in Human Annotations
  29. Preferences on a Budget: Prioritizing Document Pairs when Crowdsourcing Relevance Judgments
  30. Does Evidence from Peers Help Crowd Workers in Assessing Truthfulness?
  31. Effects of Technological Interventions for Self-regulation: A Control Experiment in Learnersourcing
  32. Hierarchical Clustering of Corals using Image Clustering
  33. An Analysis of the Australian Political Discourse in Sponsored Social Media Content
  34. On the State of Reporting in Crowdsourcing Experiments and a Checklist to Aid Current Practices
  35. Charting the Design and Analytics Agenda of Learnersourcing Systems
  36. Report on the first workshop on bias in automatic knowledge graph construction at AKBC 2020
  37. Modelling User Behavior Dynamics with Embeddings
  38. The COVID-19 Infodemic
  39. How to make crowd workers earn an hourly wage
  40. On Understanding Data Worker Interaction Behaviors
  41. Can The Crowd Identify Misinformation Objectively?
  42. Representation learning for entity type ranking
  43. Health Card Retrieval for Consumer Health Search
  44. On Transforming Relevance Scales
  45. Understanding Worker Moods and Reactions to Rejection in Crowdsourcing
  46. Quality Control Attack Schemes in Crowdsourcing
  47. Health Cards for Consumer Health Search
  48. Implicit Bias in Crowdsourced Knowledge Graphs
  49. Scalpel-CD: Leveraging Crowdsourcing and Deep Probabilistic Modeling for Debugging Noisy Training Data
  50. Deadline-Aware Fair Scheduling for Multi-Tenant Crowd-Powered Systems
  51. All Those Wasted Hours
  52. Novel insights into views towards H1N1 during the 2009 Pandemic: a thematic analysis of Twitter data
  53. Non-parametric Class Completeness Estimators for Collaborative Knowledge Graphs—The Case of Wikidata
  54. Semantic Interlinking
  55. The Impact of Task Abandonment in Crowdsourcing
  56. The Evolution of Power and Standard Wikidata Editors: Comparing Editing Behavior over Time to Predict Lifespan and Volume of Edits
  57. Can User Behaviour Sequences Reflect Perceived Novelty?
  58. Moral Panic through the Lens of Twitter
  59. Investigating User Perception of Gender Bias in Image Search
  60. On Fine-Grained Relevance Scales
  61. On the Volatility of Commercial Search Engines and its Impact on Information Retrieval Research
  62. Crowd Anatomy Beyond the Good and Bad: Behavioral Traces for Crowd Worker Modeling and Pre-selection
  63. Measuring the Effect of Public Health Campaigns on Twitter: The Case of World Autism Awareness Day
  64. Augmenting Intelligence with Humans-in-the-Loop (HumL@WWW2018) Chairs' Welcome & Organization
  65. Chapter 4: Using Twitter as a Data Source: An Overview of Ethical, Legal, and Methodological Challenges
  66. Understanding Engagement through Search Behaviour
  67. Considering Assessor Agreement in IR Evaluation
  68. Modus Operandi of Crowd Workers
  69. An Introduction to Hybrid Human-Machine Information Systems
  70. Towards building a standard dataset for Arabic keyphrase extraction evaluation
  71. Scheduling Human Intelligence Tasks in Multi-Tenant Crowd-Powered Systems
  72. Contextualized ranking of entity types based on knowledge graphs
  73. A Tutorial on Leveraging Knowledge Graphs for Web Search
  74. The Relationship Between User Perception and User Behaviour in Interactive Information Retrieval Evaluation
  75. Hybrid human–machine information systems: Challenges and opportunities
  76. Pooling-based continuous evaluation of information retrieval systems
  77. Human Beyond the Machine: Challenges and Opportunities of Microtask Crowdsourcing
  78. The Dynamics of Micro-Task Crowdsourcing
  79. Understanding Malicious Behavior in Crowdsourcing Platforms
  80. Correct Me If I'm Wrong
  81. B-hist: Entity-centric search over personal web browsing history
  82. Hippocampus
  83. Effective named entity recognition for idiosyncratic web collections
  84. Entity disambiguation in tweets leveraging user social profiles
  85. Large-scale linked data integration using probabilistic reasoning and crowdsourcing
  86. NoizCrowd: A Crowd-Based Data Gathering and Management System for Noise Level Data
  87. Ontology-Based Word Sense Disambiguation for Scientific Literature
  88. TRank: Ranking Entity Types Using the Web of Data
  89. The Bowlogna ontology: Fostering open curricula and agile knowledge bases for Europe's higher education landscape
  90. ZenCrowd
  91. BowlognaBench—Benchmarking RDF Analytics
  92. Combining inverted indices and structured search for ad-hoc object retrieval
  93. Predicting the Future Impact of News Events
  94. From people to entities
  95. Visual interfaces for stimulating exploratory search
  96. Report on INEX 2009
  97. Why finding entities in Wikipedia is difficult, sometimes
  98. Leveraging personal metadata for Desktop search: The Beagle++ system
  99. Dear search engine: what's your opinion about...?
  100. Entity summarization of news articles
  101. Exploiting click-through data for entity retrieval
  102. Overview of the INEX 2009 Entity Ranking Track
  103. Ranking Entities Using Web Search Query Logs
  104. TAER
  105. The missing links
  106. An Architecture for Finding Entities on the Web
  107. Report on INEX 2008
  108. A Vector Space Model for Ranking Entities and Its Application to Expert Search
  109. How to Trace and Revise Identities
  110. L3S at INEX 2008: Retrieving Entities Using Structured Information
  111. Overview of the INEX 2008 Entity Ranking Track
  112. A Model for Ranking Entities and Its Application to Wikipedia
  113. Social recommendations of content and metadata
  114. Leveraging semantic technologies for enterprise search
  115. A Classification of IR Effectiveness Metrics
  116. L3S at INEX 2007: Query Expansion for Entity Ranking Using a Highly Accurate Ontology
  117. Ranking Categories for Web Search
  118. Semantically Enhanced Entity Ranking