|
| 1 | +--- |
| 2 | +# static info |
| 3 | +layout: task |
| 4 | +year: 2026 |
| 5 | +hide: false |
| 6 | + |
| 7 | +# required info |
| 8 | +title: "NewsImages: Retrieval and generative AI for news thumbnails" |
| 9 | +subtitle: NewsImages |
| 10 | +blurb: "Participants receive a large set of English-language articles from international publishers. We provide a comprehensive dataset of news articles, including article title, article lead, URL (for retrieving the full text), and an editorially assigned image. Given the text of a news article, the task is to provide a fitting image. Teams are encouraged to explore different image retrieval and image generation techniques to provide article-specific visuals." |
| 11 | +--- |
| 12 | + |
| 13 | +<!-- # please respect the structure below--> |
| 14 | +*See the [MediaEval 2026 webpage](https://multimediaeval.github.io/editions/2026/) for information on how to register and participate.* |
| 15 | + |
| 16 | +#### Task description |
| 17 | + |
| 18 | +The NewsImages challenge requires participants to match news articles with images that have been retrieved or generated. |
| 19 | +Participants receive a large set of English-language articles from international publishers. We provide a comprehensive dataset of news articles, including article title, article lead, URL |
| 20 | +(for retrieving the full text), and an editorially assigned image. Given the text of a news article, the task is to provide a fitting image. Teams are encouraged to explore different image retrieval and image generation |
| 21 | +techniques to provide article-specific visuals. Pure retrieval workflows, generation workflows, or hybrid approaches are all possible. The workflows must not rely on any third-party closed-sourced APIs or resources. |
| 22 | + |
| 23 | +We are particularly interested in non-photorealistic images that do not suggest they accurately represent real events, in order to avoid misleading and/or deceiving readers. |
| 24 | + |
| 25 | +The primary criterion for evaluation is image fit, i.e., whether the images accurately capture the key attributes of the text article without depicting any important elements not present in the article. |
| 26 | +This is done in the final phase of the challenge, where participants will take part in a crowdsourced evaluation event to rate the submissions. We provide an online interface where teams will be shown images |
| 27 | +produced by their peers. |
| 28 | + |
| 29 | +Participating teams will write short working-notes papers that are published in the MediaEval Workshop Working Notes Proceedings. We welcome two types of papers: first, conventional benchmarking papers, |
| 30 | +which describe the methods that the teams use to address the task and analyze the results and, second, "Quest for Insight" papers, which address a question aimed at gaining more insight into the task, but do not |
| 31 | +necessarily present task results. Example questions for "Question for Insight" papers are below. |
| 32 | + |
| 33 | +#### Motivation and background |
| 34 | + |
| 35 | +News publishers and recommender systems depend on images and thumbnails to engage readers with news articles. Technology has advanced to the point that it is possible to automatically find a matching image |
| 36 | +(image retrieval) or generate an appropriate image (generative AI) for a news article. Although these techniques present an opportunity for the news media, they also present a great number of technical and ethical |
| 37 | +challenges. It is of critical importance that the match between the news and the image is accurate. Images should not mislead or deceive readers into assuming they represent a real-life situation. |
| 38 | + |
| 39 | +Research in multimedia and recommender systems typically assumes a straightforward relationship between images and text that co-occur. For example, in image captioning [1], the caption is often assumed to |
| 40 | +describe the literally depicted content of the image. In contrast, when images accompany news articles, the relationship becomes less clear [2]. Since images are often unavailable for the most recent news messages, |
| 41 | +stock images, archived photos, or even generated images are used as alternatives. Here, preliminary studies showed that users prefer AI-generated content over stock images [3, 4]. The goal of this task is to investigate |
| 42 | +these intricacies in more depth, in order to understand the implications that it may have for the areas of journalism and news personalization. |
| 43 | + |
| 44 | +#### Target group |
| 45 | + |
| 46 | +This task targets researchers interested in investigating the retrieval and generation of images for news, as well as the connection between images and text. This includes teams working in areas such as computer vision, |
| 47 | +recommender systems, cross-modal information retrieval, and news analysis, from both academic and industrial settings. |
| 48 | + |
| 49 | +We especially welcome interdisciplinary teams, including researchers from communication science, media studies, and computational social science (to name a few), as these perspectives are essential for understanding |
| 50 | +how images influence news perception, audience engagement and the processing by which the public makes meaning of what the information provided by the news media. |
| 51 | + |
| 52 | +#### Data |
| 53 | + |
| 54 | +The challenge uses the open-source [GDELT](https://www.gdeltproject.org/) news dataset. We create and share a subset of 8,500 news articles collected during 2022 and 2023. The news articles are all in English. Each item includes the article title, |
| 55 | +article lead, and the original image. The article text itself is not shared, but participants are free to retrieve it from the original source. We ask participants to use the |
| 56 | +[Yahoo-Flickr Creative Commons 100 Million (YFCC100M)](https://www.multimediacommons.org/) dataset as the source for images in any retrieval-based steps of their workflows. This dataset is primarily intended for testing and development. The final |
| 57 | +evaluation set may include articles from other sources (will be shared with participants shortly before the final evaluation event). Furthermore, we will provide example images that teams can use as baselines. |
| 58 | +We encourage everyone to use them for internal user testing to assess the quality of the runs. |
| 59 | + |
| 60 | +#### Evaluation methodology |
| 61 | + |
| 62 | +The generated and retrieved image submissions of the participants will be evaluated during a crowd-sourced event. To that end, the organizers provide an online tool and distribute accounts to participants. |
| 63 | +The participants are requested to rate the fit of the images submitted by their peers. The winning team is determined by the overall average rating for submitted images. In addition to the qualitative |
| 64 | +questions outlined above, the rating procedure for generated images will include a check for policy compliance. The policy applied within the scope of this challenge is based on existing AI guidelines in news media: |
| 65 | +All image submissions should adhere to current editorial standards and not suggest that they accurately represent real events, deceive, or mislead readers in any other way. |
| 66 | + |
| 67 | +Staying true to the principles of MediaEval, which promote reproducible research, the submissions of both retrieved and generated images must include the entire workflow. For image generation, we encourage |
| 68 | +participants to use tools that can automatically embed the workflow used for their generation in the image file, e.g., [ComfyUI](https://github.com/Comfy-Org/ComfyUI) or |
| 69 | +[WebUI](https://github.com/AUTOMATIC1111/stable-diffusion-webui). Furthermore, we ask participants to use open-source solutions that can be set up and run locally. |
| 70 | + |
| 71 | +#### Quest for insight |
| 72 | + |
| 73 | +This task invites participants to further reflect on the role and potential of news images in terms of generation and retrieval. The organizers encourage participants to explore further topics related to |
| 74 | +the task. You are welcome to investigate your own research in the domain of news and images. (You can do so without having prepared a challenge submission.) Possible research questions are: |
| 75 | + |
| 76 | +* Are generated images better suited than real archival images? |
| 77 | +* What parameters are most important for generating perfectly matching images? |
| 78 | +* Which categories of news are best suited for image generation/retrieval? |
| 79 | +* What do users expect from images embedded in news articles? |
| 80 | +* Do news readers care more about generated or non-generated images? |
| 81 | +* Are photographs or illustrations more effective in capturing the readers’ attention? |
| 82 | +* What are the existing policies and guidelines for retrieving/generating images, and are they in line with readers’ expectations? |
| 83 | +* What is the impact of item properties (e.g., image style or selection of subject) across the dimension of: |
| 84 | + * Trustworthiness: Does the image look like it originates from a trustworthy source? |
| 85 | + * Promoting stereotypes or Biases: Does the image promote any stereotypes or biases? |
| 86 | + * Attractiveness/clickbaitness: Does the image capture the reader’s attention? |
| 87 | + * Visual quality: Does the image look professional and not include any flaws? |
| 88 | + |
| 89 | +#### Participant information |
| 90 | + |
| 91 | +* Register as a team consisting of one, but ideally multiple people. You are free to submit to either the challenge or insight track. |
| 92 | +* The following deliverables are expected: Paper (incl. revision), presentation (incl. video), codebase (shared publicly), and written reviews of the paper submissions of your peers. |
| 93 | +* You are requested to release your code and/or workflow so that we can reproduce your submission. |
| 94 | +* Submitting your work: Details will be announced together with the dataset release. |
| 95 | +* Evaluate the event: Details will be announced once all teams have submitted their runs. |
| 96 | +* Preparing Your Working Notes Paper: Instructions on preparing your working notes paper will be available at a later date. |
| 97 | +* Please contact your task organizers with any questions on these points. |
| 98 | + |
| 99 | +#### References and recommended reading |
| 100 | + |
| 101 | +[1] Hossain, M. Z., Sohel, F., Shiratuddin, M. F., & Laga, H. (2019). [A comprehensive survey of deep learning for image captioning](https://dl.acm.org/doi/abs/10.1145/3295748). ACM Computing Surveys (CSUR), 51(6), p. 1-36. |
| 102 | + |
| 103 | +[2] Oostdijk, N., van Halteren, H., Bașar, E., & Larson, M. (2020). [The Connection between the Text and Images of News Articles: New Insights for Multimedia Analysis](https://www.aclweb.org/anthology/2020.lrec-1.535/). In Proceedings of The 12th Language Resources and Evaluation Conference, p. 4343-4351. |
| 104 | + |
| 105 | +[3] Heitz, L., Bernstein, A. & Rossetto, L. (2024). [An Empirical Exploration of Perceived Similarity between News Article Texts and Images](https://ceur-ws.org/Vol-3658/paper8.pdf). MediaEval 2023 Working Notes Proceedings. |
| 106 | + |
| 107 | +[4] Heitz, L., Chan, Y. K., Li, H., Zeng, K., Bernstein, A. & Rossetto, L. (2024). [Prompt-based Alignment of Headlines and Images Using OpenCLIP](https://ceur-ws.org/Vol-3658/paper7.pdf). MediaEval 2023 Working Notes Proceedings. |
| 108 | + |
| 109 | +#### Task organizers |
| 110 | + |
| 111 | +* Lucien Heitz, Informfully, Switzerland, lucien.heitz@informfully.ch |
| 112 | +* Bruno N. Sotic, University of Amsterdam, the Netherlands, nadalic.sotic@uva.nl |
| 113 | +* Ali Ahmadi Katamjani, Utrecht University, the Netherlands, a.ahmadikatamjani@uu.nl |
| 114 | +* Bram Bakker, Radboud University Nijmegen, the Netherlands, bram.bakker2@ru.nl |
| 115 | +* Qi Bi, Utrecht University, the Netherlands, q.bi@uu.nl |
| 116 | +* Luca Rossetto, Dublin City University, Ireland |
| 117 | + |
| 118 | +#### Task Advisors |
| 119 | + |
| 120 | +* Benjamin Kille, NTNU, Trondheim, Norway |
| 121 | +* Andreas Lommatzsch, TU Berlin, Germany |
| 122 | +* Duc-Tien Dang-Nguyen, University of Bergen, Norway |
| 123 | +* Mehdi Elahi, University of Bergen, Norway |
| 124 | +* Jaap Kamps, University of Amsterdam, The Netherlands, kamps@uva.nl |
| 125 | +* Daan Stokvis, University of Groningen, the Netherlands |
| 126 | + |
| 127 | +#### Task schedule |
| 128 | + |
| 129 | +* January: Releasing the test and development dataset. |
| 130 | +* February-April: Teams work on their submissions. |
| 131 | +* April: Sharing the evaluation dataset. |
| 132 | +* May: Online evaluation event. |
| 133 | +* May: Return of evaluation results. |
| 134 | +* June: Deadline for submitting working notes paper (incl. code/pipeline). |
| 135 | +* June: MediaEval Workshop in Amsterdam. |
| 136 | + |
| 137 | + |
0 commit comments