All Stories

  1. A Cost-Effective Framework to Evaluate LLM-Generated Relevance Judgements
  2. DP-COMET: A Differential Privacy Contextual Obfuscation MEchanism for Texts in Natural Language Processing
  3. Getting off the DIME: Dimension Pruning via Dimension Importance Estimation for Dense Information Retrieval
  4. Projection-Displacement-Based Query Performance Prediction for Embedded Space of Dense Retrievers
  5. A Comparative Study of Large Language Models and Traditional Privacy Measures to Evaluate Query Obfuscation Approaches
  6. Evaluating Multi-Dimensional Cumulated Utility in Information Retrieval
  7. LLM4Eval: Large Language Model for Evaluation in IR
  8. CoSRec: A Joint Conversational Search and Recommendation Dataset
  9. Variations in Relevance Judgments and the Shelf Life of Test Collections
  10. CoDIME: A Counterfactual Approach for Dimension Importance Estimation through Click Logs
  11. Report on the 1st Workshop on Large Language Model for Evaluation in Information Retrieval (LLM4Eval 2024) at SIGIR 2024
  12. pyPANTERA: A Python PAckage for Natural language obfuscaTion Enforcing pRivacy & Anonymization
  13. LLM4Eval: Large Language Model for Evaluation in IR
  14. Dimension Importance Estimation for Dense Information Retrieval
  15. Report on the Collab-a-Thon at ECIR 2024
  16. Report on the 13th Italian Information Retrieval Workshop (IIR 2023)
  17. Towards Query Performance Prediction for Neural Information Retrieval: Challenges and Opportunities
  18. Perspectives on Large Language Models for Relevance Judgment
  19. A Geometric Framework for Query Performance Prediction in Conversational Search
  20. Modelling and Explaining IR System Performance Towards Predictive Evaluation
  21. Report on the 1st Workshop on Query Performance Prediction and Its Evaluation in New Tasks (QPP++ 2023) at ECIR 2023