Understanding image-text relations and news values for multimodal news analysis

dc.bibliographicCitation.articleNumber1125533
dc.bibliographicCitation.journalTitleFrontiers in Artificial Intelligenceeng
dc.bibliographicCitation.volume6
dc.contributor.authorCheema, Gullal S.
dc.contributor.authorHakimov, Sherzod
dc.contributor.authorMüller-Budack, Eric
dc.contributor.authorOtto, Christian
dc.contributor.authorBateman, John A.
dc.contributor.authorEwerth, Ralph
dc.date.accessioned2024-03-06T07:21:41Z
dc.date.available2024-03-06T07:21:41Z
dc.date.issued2023
dc.description.abstractThe analysis of news dissemination is of utmost importance since the credibility of information and the identification of disinformation and misinformation affect society as a whole. Given the large amounts of news data published daily on the Web, the empirical analysis of news with regard to research questions and the detection of problematic news content on the Web require computational methods that work at scale. Today's online news are typically disseminated in a multimodal form, including various presentation modalities such as text, image, audio, and video. Recent developments in multimodal machine learning now make it possible to capture basic “descriptive” relations between modalities–such as correspondences between words and phrases, on the one hand, and corresponding visual depictions of the verbally expressed information on the other. Although such advances have enabled tremendous progress in tasks like image captioning, text-to-image generation and visual question answering, in domains such as news dissemination, there is a need to go further. In this paper, we introduce a novel framework for the computational analysis of multimodal news. We motivate a set of more complex image-text relations as well as multimodal news values based on real examples of news reports and consider their realization by computational approaches. To this end, we provide (a) an overview of existing literature from semiotics where detailed proposals have been made for taxonomies covering diverse image-text relations generalisable to any domain; (b) an overview of computational work that derives models of image-text relations from data; and (c) an overview of a particular class of news-centric attributes developed in journalism studies called news values. The result is a novel framework for multimodal news analysis that closes existing gaps in previous work while maintaining and combining the strengths of those accounts. We assess and discuss the elements of the framework with real-world examples and use cases, setting out research directions at the intersection of multimodal learning, multimodal analytics and computational social sciences that can benefit from our approach.eng
dc.description.fondsTIB_Fonds
dc.description.versionpublishedVersion
dc.identifier.urihttps://oa.tib.eu/renate/handle/123456789/14495
dc.identifier.urihttps://doi.org/10.34657/13526
dc.language.isoeng
dc.publisherLausanne : Frontiers Media
dc.relation.doihttps://doi.org/10.3389/frai.2023.1125533
dc.relation.essn2624-8212
dc.rights.licenseCC BY 4.0 Unported
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subject.ddc004
dc.subject.othercomputational analyticseng
dc.subject.otherimage-text relationseng
dc.subject.otherjournalismeng
dc.subject.othermachine learningeng
dc.subject.othermultimodalityeng
dc.subject.othernews analysiseng
dc.subject.othernews valueseng
dc.subject.othersemioticseng
dc.titleUnderstanding image-text relations and news values for multimodal news analysiseng
dc.typeArticleeng
dc.typeTexteng
tib.accessRightsopenAccess
wgl.contributorTIB
wgl.subjectInformatik
wgl.typeZeitschriftenartikel
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
5frai-06-1125533.pdf
Size:
9.73 MB
Format:
Adobe Portable Document Format
Description: