New Use Cases Demo Videos Now Live!

AI4Media has recently launched a series of seven demo videos that showcase the groundbreaking achievements of AI4Media’s use cases. These videos highlight the strong collaboration between AI researchers and media industry professionals across Europe, addressing critical challenges and exploring new market opportunities.

Each video offers a deep dive into how AI is revolutionizing media workflows and processes, providing viewers with a clear understanding of the technology’s transformative impact on the industry.

  • AI Tools for Media Verification: Insights & Demos on Use Case 1: This video introduces nine AI-driven services developed by ATC, DW, and various technical partners. These tools support media verification by offering insights, detecting fake content, and countering disinformation. Integrated into a special version of Truly Media, these services enable efficient testing and collaboration.
    [PLAY UC1 VIDEO-DEMO]
  • AI for News: The Smart News Assistant – Insights & Demos on Use Case 2: Discover how VRT is transforming journalism with AI tools that adapt stories across formats, from text and video to Instagram posts and vertical videos. Key features like frame extraction and video summarization streamline content creation, making it easier to engage audiences on multiple platforms.
    [PLAY UC2 VIDEO-DEMO]
  • Transforming Media Workflows with AI: Insights & Demos on Use Case 3: Alberto Messina from RAI’s R&D Department, CRITS discusses how AI4Media’s Use Case 3 is streamlining media workflows, particularly in covering both planned and unexpected events. The video highlights key components like the Face Management Framework and Visione System, along with the challenges of integrating AI into media operations. 
    [PLAY UC3 VIDEO-DEMO]
  • Audio Analysis in Media Research: Insights & Demos on Use Case 4: This video introduces PAM (Partial Audio Matching), developed by NISV, a tool that tracks and analyses the reuse of audio segments across media programs. PAM helps researchers understand how content, like a Covid-19 press conference, is repurposed in other broadcasts, offering insights into media dissemination and framing.
    [PLAY UC4 VIDEO-DEMO]
  • AI-Powered Music Search for Video Game Production: Insights & Demos on Use Case 5: This video introduces Music-Explorer, an AI-driven search engine developed in the AI4Media project by Fraunhofer IDMT & IRCAM to help video game producers find music that fits their video game’s ambiance. Using AI to analyze genre, mood, and similarity, Music-Explorer simplifies the music selection process. The video demonstrates how producers can use it to create audio moodboards, develop prototypes, and inspire composers.
    [PLAY UC5 VIDEO-DEMO]
  • AI-Driven Music Composition: Insights & Demos on Use Case 6: This video showcases a music composition tool developed by BSC-CNS within the AI4Media project. It allows users to create music by interacting with generative models. The demonstration shows how a simple MIDI track is expanded and converted into a final audio format through a user-friendly interface that combines multiple models on a single platform.
    [PLAY UC6 VIDEO-DEMO]
  • AI-Powered Video Content Organisation & Moderation: Insights & Demos on Use Case 7: Learn how IMAGGA’s AI technology categorizes video archives and moderates content in real-time. The video highlights the technology’s impact on media, development challenges, and future potential, showcasing how a journalist uses it to prepare reports on World War II and the 2024 Olympics by identifying key figures, symbols, and landmarks.
    [PLAY UC7 VIDEO-DEMO]

These demo videos offer valuable insights into the cutting-edge AI solutions being developed through AI4Media, illustrating how these technologies are poised to reshape the media landscape in the coming years.

The full playlist is available on the AI4Media YouTube channel [HERE].

AI4Media’s Lasting Legacy in Advancing AI for Media

Artificial Intelligence (AI) is revolutionising the media industry, enhancing content creation, analysis, and distribution, and providing deeper insights into complex social dynamics both online and offline. AI empowers the media industry to support a democratic society, enrich our lives, and foster creativity like never before. However, the deployment of AI also introduces significant ethical challenges and risks for individuals and society at large.

In response to these opportunities and challenges, the AI4Media Network of Excellence has established an extensive network of researchers across Europe and beyond, dedicated to exploring AI’s applications in media, society, and democracy. Over nearly four years, AI4Media has become a beacon of innovation, collaboration, and progress, linking AI with the media industry.

AI4Media’s Four Pillars of Success:

  1. A Dynamic Network (LINK): Comprising over 200 organisations from academia, industry, and civil society, this network fosters a collaborative environment for sharing ideas, resources, and new developments. Associate members benefit from joint projects, a vast pool of AI knowledge, and enhanced visibility through cross-organizational exchanges and access to educational materials.
  2. AI Media Observatory (LINK): Serving as a crucial resource for stakeholders, the Observatory offers up-to-date information and insights on AI in media. It identifies trends and challenges, supporting informed decision-making and fostering responsible AI development. The Observatory is useful for media professionals, AI researchers, policymakers, educators, and anyone interested in AI’s impact on media.
  3. International AI Doctoral Academy (AIDA) (LINK): Enhancing AI education across Europe, AIDA attracts PhD students and promotes academic excellence. It provides a variety of educational programs and resources, standardising AI education across European universities and supporting collaborative opportunities. AIDA has offered over 80 educational courses attended by more than 2,000 students/researchers/professionals and provides a repository with over 300 educational resources.
  4. Scientific and Research Outputs (LINK): AI4Media has made substantial contributions through publications, open datasets, software tools, and strategic documents. These resources drive innovation in AI and are vital for researchers, AI professionals, media stakeholders, policymakers, and educators. AI4Media’s research activities have yielded over 400 scientific publications, numerous reports, nearly 100 open-access software tools, and more than 40 open datasets. The project communicates its key outputs through factsheets, booklets, and infographics, collectively referred to as “results in brief.”

AI4Media’s Key Contributions

  • AI4Media Strategic Research Agenda (SRA) (LINK): The AI4Media SRA outlines strategic planning for AI research and community engagement. Initially released in March 2023 and updated in July 2024, the SRA includes new developments in Generative AI, driven by the widespread use of Large Language Models and Large Multimodal Models. It addresses significant opportunities and challenges these technologies present to media, creative industries, and society.

  • AI4Media Roadmap on AI Technologies and Applications for the Media Industry (LINK): The AI4Media Roadmap explores how AI technologies are transforming the media sector and provides insights into future developments. It examines opportunities for growth and transformation while addressing relevant risks and mitigation strategies.

  • AI4Media Junior Fellows Exchange Program (LINK): The program facilitates exchanges of early career researchers to improve their skills and knowledge in AI for the media and society. It has facilitated 88 exchanges involving 65 organisations across Europe and beyond, producing important outcomes in publications, open software, and datasets.

  • AI4Media Open Call #1 (LINK) and Open Call #2 (LINK): AI4Media launched two Open Calls targeting entrepreneurs, companies, and researchers. From 155 applications, 20 projects were funded, providing up to €50,000 in equity-free funding, coaching, and promotional services. These projects developed innovative methods and applications addressing significant challenges in media and AI.

  • AI4Media Use Cases (LINK): AI4Media implemented 7 use cases in collaboration with European media organisations. These use cases addressed disinformation, news production, media moderation, audiovisual archives, game design, artistic co-creation, and social science research. They highlight AI’s application across the media industry value chain.

  • AI4Media Technological Highlights: Four booklets showcase AI4Media’s key technological highlights:
    • Human- and Society-Centered AI – (LINK)
    • New Learning Paradigms & Distributed AI – (LINK)
    • Trustworthy AI – (LINK)
    • Content-Centered AI – (LINK)

  • Video Series “AI Applications for the Media Industry” (LINK): AI4Media produced a series of videos covering AI applications in news production, robot journalism, social media, entertainment/movie production, games, music, and publishing.

  • LLMaker: Consistent Game Content Creation (LINK): LLMaker is a specialised tool for iterative game content co-design through chat interactions. It bridges human designers and content generation systems, leveraging large language models to interpret designer requests into actionable instructions.

Through these initiatives, AI4Media has left a lasting legacy in the AI and media landscape, providing a wealth of resources, research, and educational opportunities. These efforts ensure that AI continues to advance responsibly and beneficially for all segments of society.

Announcing the Development of AI4Media Technological and Research Highlights Booklets

AI4Media is excited to announce the release of a series of booklets featuring the key technological and research highlights from the AI4Media project. These four comprehensive booklets aim to advance the next generation of core AI research for the media industry and society, enhancing Europe’s capabilities in media AI. Each booklet focuses on significant, groundbreaking, or impactful developments across different core AI research areas developed under AI4Media.

1. Booklet on Human- and Society-Centered AI – [LINK]

This booklet explores AI4Media’s contributions toward understanding the broad societal impacts and individual effects of AI. As AI technologies become increasingly prevalent in everyday applications, these insights are crucial for ensuring responsible deployment. The booklet delves into how AI can be harnessed to benefit society while addressing ethical considerations and potential risks.

2. Booklet on New Learning Paradigms & Distributed AI – [LINK]

This booklet presents innovative learning paradigms specifically tailored for media content and applications. These advancements push the boundaries of deep learning and state-of-the-art AI technologies. It highlights novel methodologies and frameworks that enhance the efficiency and effectiveness of AI systems in processing and understanding media content.

3. Booklet on Trustworthy AI – [LINK]

This booklet details new techniques and algorithms developed by AI4Media to foster trustworthy AI for media industry applications. It covers four key areas: Adversarial Robustness, Explainability and Interpretability, Privacy and Security, and Fairness. Additionally, this booklet introduces a novel AI benchmarking tool developed as part of AI4Media, which sets new standards for evaluating AI systems’ reliability and transparency.

4. Booklet on Content-Centered AI – [LINK]

This booklet presents key outcomes on the AI’s role in analysing, generating, and managing media content. It emphasizes technologies that enhance content creation and dissemination in the media sector. The insights provided in this booklet showcase how AI can revolutionize media workflows, from content production to audience engagement.

Each booklet serves as a valuable resource for stakeholders in the AI and media industries, providing in-depth insights into AI4Media’s leading-edge research and its potential applications. These booklets are designed to inform and inspire researchers, industry professionals, policymakers, and educators about the transformative potential of AI in the media sector.

By sharing these advancements, AI4Media continues to contribute to the development of responsible and impactful AI technologies that support the media industry and broader societal goals. We invite you to explore these booklets and discover the innovative work being done to shape the future of AI and media.

AI4Media’s Achievements and Impact on Society

Over the past four years, AI4Media has made significant strides in harnessing the power of artificial intelligence to address key societal challenges and enhance various aspects of public life. This comprehensive effort has resulted in notable advancements in combating disinformation, improving public understanding of AI, and developing tools that support democratic processes and personal privacy. These are some of the key achievements that highlight AI4Media’s impact on citizens and society:

“Artificial Intelligence: Possibilities and Challenges” Exhibition

AI4Media co-organised a temporary museum exhibition titled “Artificial Intelligence: Possibilities and Challenges” at the NOESIS Science Center & Technology Museum in Thessaloniki, open from April 2024. This exhibition targeted school children and the general public, aiming to demystify AI by exploring topics such as generative AI, AI bias, disinformation, and sustainability. Featuring numerous interactive elements, the exhibition provided an engaging and educational experience, helping to foster a better understanding of AI’s potential and challenges among visitors.

Advancing the Fight Against Disinformation

As part of their use case within AI4Media, Germany’s international broadcaster Deutsche Welle (DW) and the Athens Technology Centre (ATC) developed a demonstrator for testing new AI services in a media business environment. Integrated into the “lab version” of the Truly Media platform, these services included verification tools for video, audio, and text content, such as deepfake analysis and text verification. The services were tested against business requirements and contributed significantly to improving Trustworthy AI and AI compliance in media tools.

Successful Use of Deepfake Detection Service

In collaboration with the Horizon Europe project vera.ai, AI4Media developed the RINE method for synthetic image detection, now integrated into the Fake News Debunker browser plugin. This tool, used by over 130,000 journalists and fact-checkers globally, has been pivotal in flagging AI-generated images in significant events, such as the European Elections, the War in Ukraine, and the Israel-Palestine conflict. The AFP (Agence France-Presse) successfully used this service to debunk disinformation during these high-profile events.

Political Barometer: Predicting EU Election Outcomes in Greece

Developed by the Artificial Intelligence and Information Analysis Laboratory of the Department of Informatics at Aristotle University of Thessaloniki (AUTH), the Political Barometer software performs political opinion polling and election result prediction using sentiment analysis of political tweets. This innovative tool, which analyses daily tweets about political parties and integrates past election results and classical poll data, demonstrated high accuracy in predicting outcomes for the Greek parliamentary elections and the European elections of June 2024.

YDSYO App for Assessing Social Media Content Impact

The YDSYO mobile app prototype, developed under AI4Media, uses AI to provide feedback on the potential real-life effects of sharing photographs on social media. It analyses visual content from the user’s smartphone, aggregates the results, and rates these profiles in situations like job searches, loan applications, or housing searches. The app offers users control mechanisms, such as masking or deleting photos, to manage their online presence effectively. All processing is done locally on the device, ensuring user privacy and control over data.

Analysing European Local News with NLP Tools

In the AI4Media project, the Idiap Research Institute in Switzerland developed an analytical framework for local news using open-source natural language processing (NLP) tools. This framework was applied to analyse news sources at both the European and hyper-local levels. By examining local media, which plays a crucial role in maintaining community ties and addressing the crisis of trust in national media, the project highlighted the importance of local journalism in the democratic process.

Through these initiatives, AI4Media has significantly contributed to societal advancements by enhancing public understanding of AI, combating disinformation, improving political discourse analysis, and providing tools for better social media content management. These achievements underscore the project’s commitment to leveraging AI for the public good, ensuring that technological advancements benefit all segments of society.

It is time to enforce AI regulation before adding more to the mix

Now the time has come to see emerging AI regulations enforced before renegotiating and developing further regulatory initiatives. That was the key takeaway from the event titled ‘EU Vision for Media Policy in the Era of AI’ organised by KU Leuven as part of the AI4Media project in mid-June.

On June 19, 2024, regulators, researchers, practitioners, and even the Flemish minister for Brussels, Youth, Media and Poverty Reduction, Benjamin Dalle, came together in one room to discuss the transformative potential of AI in the media sector and to look forward to how regulation can make a meaningful impact in how AI is developed and used. The event was hosted at the Belgian Institute for Postal Services and Telecommunications (BIPT) in Brussels and organised by KU Leuven’s (KUL) Centre for IT & IP Law (CiTiP) as part of the Horizon2020 project AI4Media.

The day started with a keynote by Benjamin Dalle, who highlighted a two-sided role for policymakers and regulators; regulating challenges and supporting AI developments in Europe. Something that became a dominant theme throughout the day, as both media practitioners and researchers highlighted how regulation should enable responsible AI development and use, but also crucially be enforced to hinder malicious use of AI and a concentration of power.

This dual need was also evident in the policy recommendations presented during the event by Lidia Dutkiewicz, Noémie Krack from KUL, and Anna Schjøtt Hansen from the University of Amsterdam (UvA). The final recommendations will be delivered to the European Commission by the end of August. They are based on four years of research to understand the core policy needs of the media sector. Here both supportive mechanisms, such as sustainable funding schemes, and mitigating measures, such as ensuring access to APIs of large platforms, were presented as highly relevant to supporting media independence, plurality, and the maintained watchdog function of media.

However, in the final panel of the day composed of Peggy Valcke from BIPT & KUL, Renate Schroeder from the European Federation of Journalism (EFJ), Júlia Tar from Euractiv, and Tomás Dodds from Leiden University, it also became clear that one important policy need is to slow down the ongoing flow of new AI legislation, to give regulators and media organisations a chance to implement, enforce and learn from their experiences with the new legal frameworks.

Looking into a period of enforcement and learning

With the Digital Services Act (DSA), Digital Markets Act (DMA) and the AI Act all being adopted within the last few years – the coming period would be one focused on creating meaningful enforcement and learning about both the positive and negative impacts of the regulation before renegotiating.

This was also highlighted in the panel that featured three of the four media regulatory bodies of Belgium, including Bernardo Herman from the Belgian Institute for Postal Services and Telecommunications (BIPT), François Jongen from der Medienrat, and Carlo Adams from the Vlaamse Regulator voor de Media (VRM). For them, this landscape was one they were only beginning to navigate and they all underlined the importance of adequate staffing, resources, and recruitment of specialized talent across disciplines as prerequisites for efficient enforcement.

A constantly evolving AI landscape will produce new challenges

The challenges of developing and integrating AI in the media sector had been discussed earlier in the day by media practitioners including Rasa Bocyte from the Netherlands Institute for Sound & Vision (NISV), Chaja Libot from the Flemish public broadcaster (VRT), Frank Visser representing the DRAMA project and Angel Spasov from Imagga.

While they all agreed that much value is to be gained from AI when it is done right – that is exactly the tricky part; getting it right. As Rasa Bocyte noted as she started her presentation: “Integration of AI in media is not straightforward”, highlighting how media professionals try to navigate this responsibly but face many dichotomies of wanting to move fast but also being organisationally cautious and slow to protect societal values.

Rasa Boçyte from the Netherlands Institute for Sound & Vision (NISV) presenting the results from the AI4Media AI in media integration workshop

While the media practitioners welcomed a pause in the stream of new legislation, they also stressed the importance of keeping up with the new challenges that AI will continue to pose for the media sector, such as the ongoing debate around copy-righted training data.

Rasa has also introduced the AI4Media initiative How is the Media Sector Responding to Content Crawling for Model Training as a concrete initiative to help gain an overview of these challenges, which was also an area where there was agreement in the room that the EU legislation yet needs to capture the full extend of the problem.

A missed legislative opportunity?

Many media practitioners and researchers find it important to consider how we regulate AI societal risks, including worker displacement, and environmental costs. An equally important aspect is the power imbalance between big tech and media and limited support for long-term sustainable funding and upskilling. These remain only limitedly addressed. While the mantra of the day remained the call for ‘no more legislation’, it was also stressed that it would be important to revisit the newly passed acts in a few years to better address these challenges once we have learned more about their real-world effects. It is only then the legislators should consider new actions to close the enforcement gaps.

Anna Schjøtt Hansen from the University of Amsterdam (UvA) presenting six cross-cutting policy needs for the media sector

The need for interdisciplinarity & collaboration

The various panels emphasized the critical need for collaboration across stakeholders active in the media sector. It was highlighted that shaping an effective EU agenda and ensuring the responsible integration of AI in the media sector cannot be achieved in isolation. The discussions underscored that collective effort and multi-stakeholder engagement are essential to navigate the complexities and harness the full potential of AI use in the media sector. The closing remarks invited all participants to commit to ongoing dialogue and cooperation to drive forward responsible technology development and AI strategy for the media sector.

Experts’ perspective on policy & regulation

Fact box: AI4Media & find out more

The event was part of the AI4Media project, which aims to strengthen Europe’s Excellence in AI for Media, Society, and Democracy and ensure the development of ethical and trustworthy AI.

Importantly the insights generated at the event will feed into the final policy recommendations that will be published by the end of August and sent to the European Commission. In the following weeks, four blog posts that discuss where the current legislation is finding and missing its mark will also be published on the AI Media Observatory’s medium page.

To gain an overview of the work that has led up to the policy recommendations, you can find the reports, factsheets, and whitepapers that disseminate the results in this brochure.

From speculations to future applications: AI4Media publishes scenarios envisioning the future horizons for AI in media

Authors: Rasa Bocyte and Johan Oomen, Netherlands Institute for Sound & Vision

News recommender systems that put the audience in the driver’s seat, highly personalised fact-checks and automatically created podcast mixtapes – these are just a few ideas captured in the booklet Exploring AI Potential and Collaborations in the Evolving European Media Landscape. The new publication, created by the AI4Media Centre of Excellence team, presents nine scenarios that capture future innovation directions in the media sector.

The publication is a result of a speculative design process – a series of creative workshops that brought media organisations, AI developers and researchers to co-create future-oriented scenarios of using AI in the media sector. Using future thinking and design thinking approaches, experts worked in interdisciplinary teams to reflect on the current use of AI in the media sector and imagine what they would like to see in this field in the near future. 

Over fifty participants from a wide range of media backgrounds took part in this process facilitated by the Netherlands Institute for Sound & Vision and ZEZA. The scenarios are based on the extensive catalogue of AI techniques and software developed by the AI4Media project. 

Imaging and building desirable futures

But why speculative designs? As new AI techniques emerge and mature, they open possibilities for new application areas that bring societal benefits as well as introduce new risks. This prompts the industry and the research community to chart new territories: what future use cases and scenarios can we imagine for these technologies? What new business opportunities will they create for the media sector? How can they learn from recent misuse of AI-driven applications and create applications that support public values? By crafting speculative scenarios, we can foresee potential risks and choose alternative paths.

The exercise of speculative design can also help with the persistent challenge of bridging the gap between fundamental AI research and practical applications. It helps researchers and industry representatives get a better understanding of each other’s perspectives and tailor the development of AI-based solutions towards specific societal needs and challenges.

Illustrations Illustrator: Magda Rysuje

Illustrator: Magda Rysuje

The scenarios presented in the booklet are designed to spark inspiration among researchers and media professionals, encouraging them to bridge the gap between cutting-edge AI research and the pressing challenges confronting the European media sector. To facilitate this, each scenario is complemented by links to relevant technologies developed by AI4Media partners. While these scenarios are not immediate market-ready solutions, they serve as a vision of the future potential of technologies in the AI4Media ecosystem.

Embedded in AI4Media

The publication’s ambitions are aligned with the overall strategy of the AI4Media Centre of Excellence (CoE). Funded by the Horizon 2020 Programme, this ambitious initiative delivers next-generation AI research and training at the service of media, society and democracy. The CoE brings together researchers, developers and media professionals from across the EU to translate fundamental AI research into concrete solutions to support the media industry value chains now and in the future. 

Over the past years, AI4Media explored seven use cases that cover a variety of topics such as disinformation, news research and production, media moderation, organisation of audiovisual archives, game design, human-machine artistic co-creation, and social sciences research. The AI4Media Observatory supports multidisciplinary community collaboration and fosters broader discussions on AI’s development, impact, and societal implications.

A call for feedback

AI4Media seeks collaboration with media partners and small to medium enterprises for further testing of the technologies presented in the booklet and to gather feedback from media professionals. If you’re interested in collaborating, please contact us at info@ai4media.eu

AI4Media Co-Organises “Artificial Intelligence: Possibilities and Challenges” Exhibition at Museum NOESIS Science Center

AI4Media, in collaboration with the NOESIS Science Center & Technology Museum in Thessaloniki, has co-organised a groundbreaking temporary exhibition titled “Artificial Intelligence: Possibilities and Challenges,” which opened in April 2024. This exhibition is designed to engage both school children and the general public, aiming to demystify the complex world of artificial intelligence.


Source: NOESIS Science Center & Technology Museum

The exhibition delves into several critical aspects of AI, including:

  • Generative AI: Demonstrating how AI can create new content, from images to text, and the implications of these technologies.
  • AI Bias: Highlighting the issues of bias in AI systems, explaining how it occurs and its impact on society.
  • Disinformation: Addressing the role of AI in spreading and combating disinformation, a growing concern in the digital age.
  • Sustainability: Exploring how AI can contribute to sustainable practices and the ethical considerations involved.

The “Artificial Intelligence: Possibilities and Challenges” exhibition is rich with interactive elements, making it an engaging educational experience. Visitors can participate in hands-on activities that illustrate how AI works and its applications in various fields. These interactive components are designed to foster a deeper understanding of AI’s potential and the challenges it presents.


Source: NOESIS Science Center & Technology Museum

By targeting school children and the general public, the exhibition aims to build a foundational understanding of AI. It strives to dispel myths and provide a balanced view of AI technologies, highlighting both their benefits and their limitations. This initiative is part of AI4Media’s broader mission to promote awareness and informed discourse around artificial intelligence.

Visit the Exhibition

For more information about the exhibition and to plan your visit, please check the information at the NOESIS Science Center & Technology Museum’s website: Artificial Intelligence: Possibilities and Challenges.

Through this exhibition, AI4Media and NOESIS are making significant strides in educating the public about AI, ensuring that visitors leave with a greater appreciation of the possibilities and challenges that this transformative technology brings.

Exploring real-world AI applications in Media: A look at seven use cases

Seven use cases have been defined by AI4Media’s industry partners, informed by emerging market opportunities and urgent industry challenges, raising specific requirements and research questions. AI4Media use cases highlight how AI applies throughout the media industry value chain, from research and content creation to production, distribution, consumption/interaction, performance and quality measurement. These industry cases play a key role in exploiting and sustaining results of AI4Media research activities. Have a look at them:

  • AI for Social Media and Against Disinformation 

This use case from Deutsche Welle (DW) and Athens Technology Center (ATC) leverages AI technologies to improve support tools used by journalists and fact-checking experts for digital content verification and disinformation detection. While partner DW provides journalistic and media-focused requirements, ATC is responsible for AI component integration and the operation of the demonstrators, Truly Media – a web-based platform for collaborative verification – and TruthNest – a Twitter analytics and bot detection tool. Two main topics are covered within the use case: 1) verification of content from social media with a focus on synthetic media detection, and 2) detection of communication narratives and patterns related to disinformation. The key motivation behind this work is to demonstrate how advanced AI support functions can enable news journalists to keep up with rapid new developments in the area of manipulated social media content, synthetic media, and disinformation.

To that end, related AI technologies that are being integrated into the use case demonstrators support journalists in detecting manipulated and synthetically generated images, videos, and audio, as well as detecting bot-generated tweets and managing their content through media summarization technologies. These AI-based tools are being developed by some of the largest research centres in Europe, such as the Centre for Research and Technology Hellas (CERTH), Fraunhofer, and CEA. We are also experimenting with AI at the edge applications for journalism. We are experimenting with how latest advances in the area can be leveraged, with a view to perform on-device critical media processing tasks, such as deepfake detection and face anonymization or NLP-based text analysis and question answering. This is considered a valuable capability in the context of counteracting disinformation, especially in cases where the media content of interest is of confidential or sensitive nature or in cases where the surrounding context does not allow it to be shared over public communication networks (e.g. areas without high-bandwidth connectivity or under strict monitoring by local authorities).

Another key aspect is the exploration of Trustworthy AI in relation to these topics and the specific needs of media organisations. Our goals are to explore and demonstrate how an AI component from a third-party provider can be enhanced in terms of transparency and robustness, to develop related AI transparency information documents for different target groups within a media organisation, and to make such transparency information available in the user interface of the demonstrators.

If you wish to know more about this work, please contact Danae Tsabouraki at d.tsabouraki@atc.gr

  • AI for News: The Smart News Assistant 

Journalists face a challenging environment where the amount of incoming content is ever increasing, while the need to publish news as fast as possible is extremely pressing. At the same time, journalists need to ensure the published content is both relevant to their audience and is a trustworthy source of information, avoiding errors and misinformation. This use case from the Flemish Public Broadcaster (VRT) focuses on interweaving smart (AI-powered) tools into day-to-day journalistic workflows in modern newsrooms, aiming to optimise repetitive tasks and create opportunities for new story formats supported by these tools. VRT is creating a Smart News Assistant, i.e., a multi-functional and AI-driven toolbox that will support journalists in monitoring, fact-checking, and creating engaging news formats.

Current work focused on investigating the workflow of a journalist and how to customise it with AI. We enhanced the Image Verification Tool from CERTH by creating a new user interface that provides step-by-step guidance through the image verification process. We developed a new prototype called Video Curator that matches incoming audiovisual content with news-related text written by journalists, in order to suggest suitable video output. Work is underway to create a new prototype that will help journalists to better understand and use data in their stories.

  • AI for Vision: Video Production and Content Automation 

Content preservation, high-quality production and process automation are at the core of the current transformation of Public Service Media (PSM) from its traditional business to the modern digital era. Emerging AI-based technologies can support PSMs in this transition by providing capabilities to simplify and accelerate content production processes and to enhance existing content, such as broadcasters’ archives. 

The use case defined by Rai, the Italian public broadcaster, focuses on three main tasks usually accomplished during everyday operations, namely content access, content production and content monitoring. Content access includes tools supporting users to find content according to specific semantic features, like persons’ names, places and organisations referenced in texts, recognising monuments depicted in images or identifying TV celebrities appearing in videos. Content production involves activities aimed at the creation and enhancement of content (e.g., video super resolution, video denoising). Content monitoring comprises some of the pillars of public media services, such as diversity analysis, content reliability assessment and social media analysis.  

The use case aims at exploring the plethora of new AI-driven tools to find the most suitable for each of these domains of application, identifying an as smooth as possible integration of each component into well established media workflows. Content access tasks have already been tackled, working on the possible introduction in the production workflows of four AI-driven components related to informative content and archive exploitation.  

Indeed, since being able to leverage on visual features instead of using only textual metadata could be of help to journalists for their search and retrieval activities, we worked on technologies allowing professionals to identify faces of TV personalities and geographic landmarks in video, as well as improving their possibilities to search for content using images as query. Another important feature that has been integrated into Rai’s tools for journalists is an AI-driven NER working on English and German content, which will improve some daily workflow of professionals working in bilingual regions of Italy. 

As for the content production activities, Public Service Media are extremely interested in video enhancement tasks, to be able to upgrade the big amount of archived content from e.g., HD to 4k (or even from SD to HD sometimes) for their possible reuse. Following this path, we assessed a super-resolution component and compared its performances with SOTA technologies obtaining promising results. Further tests will follow using different models. Content monitoring activities will also be tackled in the next period. 

  • AI for Social Sciences and Humanities 

Researchers working in media studies, history, political sciences, and other fields within social sciences and humanities (SSH) have greatly benefited from the digitization of audiovisual archives. It has expanded the scale and scope of their investigations, stimulating new research questions – a great example of this is an examination of political party representation in the media during the election period. The Netherlands Institute for Sound & Vision developed a use case that investigates how AI-based tooling could enhance SSH research with big data from archival collections. Specifically, AI4Media has provided us with an opportunity to expand the capabilities of the CLARIAH Media Suite, a digital environment for conducting research with multimodal datasets from Dutch heritage organisations. 

Over the last two years we have been collaborating with Fraunhofer IDMT to develop the Partial Audio Functionality (PAM) for the Media Suite that allows researchers to detect and trace the reuse of audiovisual programs based on the matching of identical audio signals. This can show ways in which moving images have been reused to frame a topic by giving source material a different meaning in a new context. For instance, a Media Suite user might choose a particular press conference and perform a PAM analysis to identify how segments from this program have been quoted in the evening news in weeks that follow, allowing them to compare how the same topic is reported on by different TV channels.

We have already performed an initial evaluation of PAM with researchers in the field of media studies. They confirmed the usefulness of the tool in studying the circulation and ‘canonization’ of images and speeches. Researchers were particularly excited to see a tool that is based on audio rather than visual analysis. This opens up new possibilities for currently underrepresented research areas, such as the analysis of soundscapes. What also became evident during this evaluation is that researchers place a high priority on the explainability and trustworthiness of AI tools. They need to be transparent about the limitations of their methods or potential biases and make their research replicable. Therefore, the next step in our work will be extending PAM with a model card based on IBM’s AI Fairness 360.

  • AI for Games: 

Digital games are one of the fastest-growing multimedia sectors with a projected market growth of $200 billion by 2023. This incredible trajectory is partly supported by a “games-as-a-service” business model, in which games are continuously developed and expanded beyond their initial release. While the steady flow of content helps with customer retention, it also puts pressure on developers because this content has to be tested and optimised before release. Artificial Intelligence (AI) can provide a radically new approach to game development and testing by allowing developers to test thousands of different configurations. AI can replace or augment existing practices by providing product evaluations faster than current methods and can evaluate potential products with a reduced need for human labour or data.

Automated Testing for Games: In the first sub-use case, Automated Testing for Games, MODL.AI demonstrates how AI tools can enhance the development process through automated testing and bug finding. The first objective is to provide a prototype of the platform where users can investigate a number of quality assurance reports generated by an AI agent. These reports are generated by a quality diversity agent run in a simple game demo. We are currently working to expand the prototype of the platform into a fully functional one, where the user can investigate a number of quality assurance reports generated by an AI agent in any game, supported by a plug-in for the world’s most prolific game engines, Unity and Unreal Engine, for easy integration for game developers.

Improved Music Analysis for Games: Even if video game producers usually ask human musicians to compose original background music, the development team needs audio examples that match the ambiance of the game in order to define audio mood boards and to provide music examples to facilitate the communication with the composers. The finding of suitable music examples is not a simple task, and it can take a long time. In this context IRCAM intends to demonstrate the benefit of AI methods in the context of developing video games. Based on an automatic analysis of music files, the demonstrator proposes an exploration of a wide music catalogue that is not manually annotated. 

In the current release, a catalogue of 105,000 songs was analysed to predict attributes (music genres, moods, tempo, etc.) and to compute similarity embeddings. Then the “Music Explorer” demonstrator – a web service – allows the exploration of the catalogue in two ways: first, the user defines the musical attributes which fit the ambiance of the game, and the service proposes a list of songs which fit the attributes. Contrarily to similar tools, here the criterion is based on automatically estimated attributes, and the method is then applicable even for catalogues which are not manually annotated. The second search method is based on music similarity. Here the user chooses a reference song, and he selects one or more music concepts (among: genre, mood, instrumentation, era, harmony and rhythm) in order to define the meaning of “similarity”. Then, the service returns the list of the closest songs in the catalogue. The analyses, for attributes and similarity search, are based on AI methods, and the web service is composed of a GUI displayed in the web browser of the user, and a back-end integrated on a distant server and running the AI components remotely. During the first evaluation, the Music Explorer demonstrator proved its usefulness and its ability to quickly find music examples, in order to help video game producers during the creation of a game.

  • AI for Human Co-Creation

This use case developed by the Barcelona Supercomputing Center explores the relationship between Human creation and AI tools for music composition. Labelled as Human co-creation, it potentially may have a deep impact on an industry feeding content to a society continuously consuming media production. We are currently developing novel tools that may contribute to an efficient creation task using AI tools, where the efforts of the artist or creator are focused on deeply creative tasks, relying on the assistant to perform less critical parts transparently during content co-creation. As the functionalities of these models can be complex to handle, the purpose is to provide to the final user – typically a music creator – a collection of well-organised functionalities and capabilities for AI-assisted music creation. These functionalities enable users to a) train and manipulate the model using a defined dataset selected by them, b) generate from the trained model novel content based on a small audio seed, and c) assess the quality of the generated audio content and publish the content on popular audio platforms

The current developments allow a non expert user to use advanced, pre-trained generative models or to prepare datasets for training under controlled conditions. We include a number of generative models released under the AI4Media project but also elsewhere. In addition, we have explored user requirements to understand the needs of a community of non experts approaching AI tools. The implementation of musical processing tools opens the possibility to create in a transparent manner content used in multiple formats. Composers use large datasets with music fragments and combine them using Machine Learning methods. While a single training may provide a large amount of content (different audio files), using different datasets improves the quality and variability of the generated output. However, the computational requirements are large and better training methods and data models are needed.

  • AI for (Re-)organisation and Content Moderation

Media companies have accumulated vast digital archives and collections of images and videos over the years. Since these collections have been gradually and iteratively built over time, often by different departments and units of media companies, they usually have little or no metadata such as tags, categories, and other types of annotations. This lack of coherent media asset organisation tailored to the media company business and services precludes the easy reusage and the successful monetisation of these media assets, and the creation and offering of new services. In addition, both big traditional media companies and more so digital media platforms combine in their collections both media content, created by these companies, but increasingly also user-generated content (UGC). Such hybrid media archives need advanced content moderation (CM) solutions, often working in real time to safeguard viewers and meet law and regulation requirements of various jurisdictions.

Currently our work focuses on including the integration and the use of Imagga’s content moderation and facial recognition technologies. Imagga has implemented novel methodologies based on advanced deep learning techniques such as CNNs and RNNs aimed at photo and video moderation – tagging, categorisation, and facial recognition As part of the content moderation, we have included object detection of infamous symbols and analysis whether the video has not-safe-for-work (NSFW) or explicit content. For facial recognition, we have included a celebrity recognition model able to recognize around 3,000 different celebrities. For each scene in each video, we have generated annotation metadata that is used for filtering and searching. The videos are split by keyframes and then processed by Imagga’s technologies to receive coordinates for infamous symbols and celebrities, present in the extracted keyframe images. These frames are also analysed for the presence of NSFW content. Then through a user-friendly web UI, the content can be searched and filtered.

Conclusions

The active guidance provided by use case partners to research partners throughout the integration process plays a crucial role in achieving success in all cases. This emphasises the significance of industry and research collaboration right from the project’s inception, highlighting that a lab-to-market transfer process requires their joint efforts. Moreover, the direct involvement of end-users in iterative and agile development processes further amplifies the potential market adoption of AI-related innovations, fostering a user-centric approach and ensuring the practical relevance of the developed solutions.

Nevertheless, as in the case of every innovation process, various challenges arise and often need to be tackled impromptu. Even more so, since we are dealing with the fast-evolving and frequently disrupted domain of digital technologies. Among these challenges are problems related to  structural and organisational differences among the consortium partners, integration complexities, usability and understandability issues, human-AI collaboration challenges, and dataset creation concerns. Moreover, ChatGPT’s public release was certainly a game changer for AI-driven innovation, although its long term impact on the media industry remains to be seen. 

To address these challenges, we sought to enhance collaboration by establishing closer, one-to-one relationships between industrial and research partners. Knowledge exchange, co-design activities, and joint events were also used to strengthen collaboration. Efforts to design more user-friendly interfaces and increase transparency for end users of the demonstrators were made in order to address usability issues. Human-AI collaboration aspects were improved through the development of transparency information and feedback mechanisms to enhance user trust in AI-generated results. Finally, careful dataset curation and Non-Disclosure-Agreements addressed bias and privacy concerns. Overall, our experience shows that a collaborative approach and ongoing adaptability are key to addressing challenges and ensuring the successful integration of AI research innovations into real-world applications.

Author(s): DanaeTsabouraki (Athens Technology Center); Birgit Gray (Deutsche Welle); Chaja Libot (VRT); Maurizio Montagnuolo (RAI); Rasa Bocyte (Netherlands Institute for Sound & Vision); Christoffer Holmgård (modl.ai); Rémi Mignot (IRCAM); Artur Garcia (BSC); Chris Georgiev (Imagga Technologies).

 

Exploring the future of Media: AI4Media’s fascinating video series

Unveiling the AI4Media Video Series

AI4Media, a European funded project at the intersection of AI and media, has curated a video series that provides a fascinating glimpse into the realm of AI applications within the media industry. Accessible to a global audience, this series aims to demystify AI’s role in shaping the future of media while highlighting the practical implications and potential of this powerful technology.

What makes the AI4Media video series truly captivating is its multifaceted exploration of AI’s applications in media. Each episode delves into a specific facet of this dynamic relationship, offering valuable insights and real-world examples. These are the exciting topics covered in the series:

    1. AI for News Production: This video highlights how AI enhances journalism, aiding in effective news reporting, especially in challenging scenarios, by optimising bandwidth, content management, and enabling real-time mapping and 3D visualisations.
    2. Robot Journalism: This video exemplifies how AI streamlines event coverage and content generation by managing extensive data, integrating information efficiently, and improving the quality of automated content, all while preserving editorial control.
    3. AI for the Next-Gen of Social Media: This video explores the various applications of AI in social media, including automating trend detection, aggregation, categorisation, analysing public sentiment and perceptions, content translation, and more.
    4. AI for Entertainment/Movie Production: This video demonstrates how AI technologies streamline the filming process, enhance content reach, and provide creative options, all while saving valuable time and resources.
    5. AI for Games: This video illustrates how AI assists medium-sized game development companies by streamlining testing processes, pinpointing issues in new content, and elevating overall productivity and quality, establishing itself as an invaluable solution in the industry.
    6. AI for Music: This video demonstrates how AI can enhance music composition and live performances, providing synchronisation support for movie soundtracks and empowering DJs with dynamic, style-adaptive music creation during live shows.
    7. AI for Publishing: This video explores how AI-driven co-creation platforms are revolutionising manuscript selection for publishers through user feedback analysis and content feature assessment.

Real-world applications and case studies are highlighted, illustrating how AI is being harnessed to address challenges and unlock new opportunities in the media ecosystem.

AI4Media’s video series goes beyond mere dissemination; it aims to empower knowledge and foster dialogue. By presenting complex concepts in an accessible manner, the series invites viewers to join the conversation surrounding AI in the media. Whether you are an industry professional, a curious enthusiast, or an academic, the series provides a platform for understanding, discussion, and engagement.

Access the Series Today

The AI4Media video series is just a click away on YouTube, accessible to anyone with an internet connection and a thirst for knowledge. To embark on this enlightening journey and explore the applications of AI and media, follow this link: AI4Media Video Series on YouTube.

Don’t miss the opportunity to uncover the transformative potential of AI in the media industry.

Author: Candela Bravo (LOBA)

Explore the new Scientific Papers page

Navigating the vast landscape of scientific papers can be daunting, but our new and improved filtering system empowers you to effortlessly refine your search. Whether you’re a student seeking cutting-edge studies or a seasoned researcher exploring AI4Media’s subjects, our revamped platform allows you to filter and customise your search results with precision. Check out the updated scientific papers’ page.

We are thrilled to introduce the new set of filters that empower you to precisely tailor your scientific paper searches:

  • Terms: The foundation of your search, allowing you to input specific keywords and phrases to pinpoint exactly what you’re looking for. Whether it’s machine learning, image detection, or the application potential of AI for the Media industry, our term filter ensures your search is laser-focused.
  • Author: Seek papers authored by your favourite experts or discover new voices in your field of interest. With this filter, you can find research directly from those who inspire you.
  • Year of Publication: Stay up-to-date with the latest research or delve into historical archives by narrowing your search to a particular publication year.
  • Institution: Explore papers affiliated with prestigious institutions or uncover hidden gems from lesser-known research centres.
  • Type of Publication: Are you searching for journal articles, conference papers, or books? Choose the publication type that suits your needs.
  • Publisher: Identify papers from trusted publishers, ensuring the credibility and quality of your sources.
  • Access Type: Make sure you access a wealth of research freely and without restrictions by filtering the open-access scientific papers.

Discovering knowledge should be intuitive, and our commitment to innovation ensures that your experience is both seamless and enriching. Welcome to the future of scientific exploration, where finding the research you seek is as simple as a few clicks. Explore, learn, and thrive with us as we continue to advance the way you access scientific papers.

Author: Mariana Carola

Unveiling propaganda on news articles: Cutting-edge models with linguistic and argumentative features

Propaganda has long been employed as a powerful communication tool to promote a cause or a viewpoint, especially in politics, despite its often misleading and harmful nature. Given the number of propagandist and fallacious messages posted on online social media everyday, the need to automatically detect and categorise propagandist content is crucial to safeguard society from its potential harm. We proposed text models that tackle these tasks and analyze the features that characterise propagandist messages. We based our proposed models on state-of-the-art transformer-based architectures and enrich them with a set of linguistic features ranging from sentiment and emotion to argumentation features. The experiments were conducted on two standard benchmarks in the Natural Language Processing field: NLP4IF’ 19 and SemEval’20-Task 11. Both are collections of news articles annotated with propaganda classes. Our models outperformed state-of-the-art systems on many of the propaganda detection and classification tasks. F1 scores of 0.72 and 0.68 were achieved on the sentence-level binary classification task for NLP4IF’ 19 and SemEval’20-Task 11 respectively. For the fragment-level classification task, our models outperformed the SOTA model in some propaganda classes. For instance, using NLP4IF’ 19, F1 scores of 0.61, 0.42 and 0.40 were obtained for “flag-waving”, “loaded language” and “appear to fear” respectively.

 

Semantic and argumentative features behind propaganda

In our pursuit to understand propaganda’s linguistic characteristics, we considered four groups of features that have previously shown links to propaganda: persuasion, sentiment, message simplicity, and argumentative features. In the persuasion group, we examined speech style, concreteness, subjectivity, and lexical complexity. For sentiment, we gathered sentiment labels, emotion labels, VAD scores, connotation, and politeness measurements. Message simplicity was analyzed through exaggeration and various text length-related metrics. To measure most of these variables we used, or constructed, a variety of lexicons. Finally, we trained classifiers that helped us extract argumentative features. That is, which parts of the text correspond to claims, premises, or none of them. This is important to understand the logical structure behind propaganda.

Propaganda’s levels of detection.

We addressed both Sentence-Level Classification (SLC), which asks to predict whether a sentence contains at least one propaganda technique, and Fragment-Level Classification (FLC), which asks to identify both the spans and the type of propaganda technique. The evaluation of the FLC task varied depending on the dataset being used. One of the main differences lies in the number of propaganda categories considered in each corpus: 18 in NLP4IF’ 19, and 14 in SemEval’20-Task 11.

Sentence-Level Classification

To tackle SLC, we employ a range of models, including BERT, T5, Linear-Neuron Attention BER, Multi-granularity BERT, BERT combined with BiLSTM, and BERT combined with logistic regression. In our proposed models, we utilize the last 3 architectures and modify them to include semantic and argumentative features. Our proposed models surpassed the state-of-the-art architectures. In some cases, semantic features alone demonstrated slightly better results than combining them with argumentation features.

Fragment-Level Classification

On the NLP4IF’19 Dataset, we evaluate various models, such as BERT, RoBERTa and the transformer-based winner architecture from the NLP4IF’19 shared task. Our proposed architectures used BERT with CRF output layers, outperforming the state-of-the-art model for several propaganda techniques.

In the SemEval’20 T11 dataset, we implement solutions based on BERT, RoBERTa, and the winning approach of the SemEval’20 T11 challenge. Our proposed model combined a transformer architecture with a BiLSTM. In addition to textual input, we fed the model with semantic and argumentation features. Also, we used a joint loss function that considers the loss at the sentence level, span level, and for the additional features. Such a model outperformed the SOTA model in some propaganda classes. In general, we noticed that using different training epochs help to detect different propaganda techniques. For instance, the classes “bandwagon and reductio ad hitlerium” and “thought-terminating cliches” are learnt best at low training epochs, while “casual oversimplification”, is learnt at high training epochs.

This task remains challenging, in particular regarding the fine-grained classification of the different propaganda classes.

What’s next?

Propaganda leverages emotional and logical fallacies and it is present in all kinds of media. That is why we have turned our attention to the study of fallacies in Twitter (now X), the bustling hub of information and opinions. This is a challenging task since fallacy identification, many times, relies on the context in which the text exists. Given the short length of tweets, such context is not always available. We are currently working on the definition of transformer-based architectures that will help us classify fallacies in this social media and continue our journey to fight misinformation and promote a more informed society.

 

Author: Mariana Chaves (UCA-3IA)

How did the European press treat the covid-19 “no-vax” phenomenon?

In AI4Media, Idiap has worked on the analysis of newspapers in different countries and their relationship with misinformation in the context of Covid-19 vaccination news.

Initially, a dataset was created, comprising more than 50,000 articles on Covid-19 vaccination from 19 newspapers across 5 European countries. From this dataset, a set of subtopics (within the main topic of covid vaccination news) was identified using topic models. Companies, countries, and individuals most frequently mentioned in each country were also identified using Named Entity Recognition techniques, while the sentiment of both headlines and full articles was determined. The results revealed consistencies across countries and subtopics (e.g. a prevalence of a neutral tone, and relatively more negative sentiment in non-neutral articles, with few exceptions like the case of vaccine brands). Moreover, distinctly high negative-to-positive sentiment ratios were identified for the “no-vax” subtopic, showing that this issue had a notably negative tone. This dataset and the results of the analysis were presented at the ACM International Workshop on Multimedia AI against Disinformation (MAD’22) in Newark, US. For more details, the paper can be consulted here.

Subsequently, Idiap directed its focus towards the “no-vax” movement theme. This line of research examines how the European press addressed reactions against the Covid-19 vaccine and the disinformation and misinformation associated with this movement. Based on the Covid-19 vaccination news dataset, Idiap employed a number of methods, including named entity recognition, word embeddings, and semantic networks, to comprehend the coverage provided by the European press within the disinformation ecosystem. The results of this multi-faceted analysis demonstrate that the European press actively countered a variety of hoaxes primarily propagated on social media and criticized the anti-vax trend, irrespective of the political orientation of the newspaper. This confirms the significance of studying the role of high-quality press in the disinformation ecosystem. This research was presented at the ACM International Workshop on Multimedia AI against Disinformation (MAD’23) in Thessaloniki, Greece. For more details, the paper can be found here

Overall, Idiap’s work serves as a point of comparison with other news sources on a topic where disinformation and misinformation have led to increased risks and negative outcomes for people’s health. We believe that linguistic analyses of high-quality press in Europe can contribute to informing the design of tools against disinformation, serving as a benchmark for what constitutes reliable information.

Authors: David Alonso del Barrio & Daniel Gatica-Perez (Idiap Research Institute)

Making synthetic image detection practical

Detecting whether an image posted on the Internet is authentic or generated by one of the recently produced generative AI models poses a significant challenge for journalists and fact checkers on a daily basis. While most people are familiar with tools and models such as Midjourney, DALL-E 2 and Stable Diffusion, there is nowadays a growing number of tools, services and apps that make synthetic image generation extremely accessible and enable anyone to create highly realistic images using plain text descriptions, which are widely known as prompts. It’s only natural that such potential can be exploited by malicious actors to spread disinformation. Therefore having capable tools in place to detect whether a suspicious image is AI-generated or not holds value for media organisations and newsrooms.

Such detection tools are also abundant, with the large majority being based on “deep learning models” – very large neural networks that have been trained to distinguish between authentic and synthetic media. In academic papers, these tools have often demonstrated to perform exceptionally well on separating between authentic and synthetic imagery. However, deploying such tools in operational settings presents several challenges.

A primary challenge is these models’ tendency to perform well in a restricted set of cases (referred to as “domain”) used for their training.  Consider a scenario where the researcher primarily used synthetic and authentic images of human faces to train the model. If a journalist wants to use this model to detect whether an image depicting a building is synthetic or not, the AI model is likely to give an unreliable response due to the domain mismatch between training (human faces) and testing (buildings). The Multimedia Knowledge and Social Media Analytics Lab (MKLab) at CERTH have recently developed a method to alleviate this issue. The method achieves better generalisation ability across different domains by training the detection model solely using high-quality synthetic images. This compels the model to “learn” quality-related artifacts instead of content-related cues. This method was presented at the international workshop on Multimedia AI against Disinformation (MAD’23) in Thessaloniki, Greece. A paper providing technical details is available as part of the workshop proceedings.

A second challenge when employing synthetic image detection models in practice is that most, if not all, available tools are in the form of web services that send the provided images to a server for analysis. This is often due to the computational intensity of the detection models, necessitating a powerful server for quick calculations. However, there are situations where journalists, fact-checkers, or citizens might be uncomfortable or at risk when sharing suspicious images with third-party services. To address this challenge, the MKLab team at CERTH leveraged a newly proposed method to “compress” detection models into a much smaller size, enabling execution on a standard smartphone.  This approach allows for  deepfake detection analysis without submitting the suspicious image to a third party. This compression uses “knowledge distillation”, where a computationally expensive model acts as a “teacher” to train a lighter model (the “student”). In experiments, the model size could be halved while maintaining nearly the same detection accuracy.  Even a 10-fold reduction was possible with only a slight decrease in accuracy. The method used for these results has been submitted for publication in an international journal, and a preprint is publicly available. 

It’s important to note a key limitation of the above results. Both focus on detecting GAN-generated images (well-known cases of such images are generated by https://thispersondoesnotexist.com/). Currently, detecting images produced by models like DALL-E 2, Stable Diffusion, and Midjourney is not feasible, although ongoing experiments show promise in developing tools that could enhance journalists’ and fact-checkers’ capabilities in countering disinformation.

Author: Akis Papadopoulos (CERTH)

AIDA: Maximising efforts toward accessible AI education and research

AIDA (International AI Doctoral Academy) is a non-profit organization comprising academic and industrial partners. It receives support from the European Networks of Excellence AI4Media, ELISE, TAILOR, HumanE-AI NET, and VISION CSA. AIDA’s primary aim is to enhance accessibility to AI education and research.

Its key objectives include:

  • Coordinating educational and training activities in AI for PhD and postdoc students among AIDA partners.
  • Establishing itself as a global reference point for all matters related to AI education and research.
  • Developing mechanisms for the sharing of educational resources in the field of PhD-level AI across universities.
  • Paving the way for future efforts aimed at creating a charter for European universities to share, accredit, and recognize PhD education credits in AI.

In line with these objectives, AIDA has undertaken significant efforts to maximize the impact and user engagement in AI education and research excellence.

More concretely, AIDA’s vision is to:

  • Cultivate a new generation of AI talents in Europe.
  • Establish itself as a leading reference in AI education.
  • Operate with a focus on realism and ensure long-term sustainability.

In this context, AIDA is a strong advocate for providing free access to AI educational resources and materials. The expansion of its AI offerings can be summarized as follows:

These figures reflect significant growth due to AIDA’s effective communication and dissemination practices. Depending on the context, AIDA selectively or publicly communicates and disseminates its efforts and directs offerings to relevant groups of recipients. This is primarily achieved through the use of social media and mailing lists.

Focusing on the user, the primary beneficiary of free access to AI education and research, AIDA has improved and enhanced its website to promote transparency, attractiveness, user-friendliness, and content enrichment.

In summary, these efforts have maximized the impact and user engagement of AIDA’s offerings, bringing it closer to establishing itself as an authority and a one-stop-shop for AI excellence in Europe.

For interested readers, becoming an AIDA member is possible by following the links below:

Author: George Bouchagiar (AUTH)

The ten projects from AI4Media’s second funding program are introducing fresh AI research and innovation for the media industry

The objective of AI4Media – Open Call #2, much like the first open call, was to engage companies and researchers in developing new research and applications for AI and media, thus contributing to the enrichment of the technological tools developed within the AI4Media network. Applicants were required to address specific challenges outlined by AI4Media partners, all of which are aligned with the Roadmap on AI technologies and applications for the Media Industry. 

Out of a total of 95 submissions received from 24 countries during the competitive open call, 10 projects were selected. The open call ran from September 29 to November 30, 2022. Eligible submissions underwent external evaluation by independent experts, and a selected group of proposals advanced to the interview stage. Each project has been awarded a grant of up to €50,000 to implement their work plan.

Throughout the remainder of the funding program, AI4Media will provide beneficiaries with tailored coaching, business support, and external visibility. Additionally, a boot camp featuring various workshops will be conducted later in 2023.

Here’s a brief overview of the funded projects and their objectives:

APPLICATION projects:

  1. JECT-CLONE (JECT.AI Limited, SME from the UK): Delivering new computational creativity capabilities as a software-as-a-service (SaaS) that autonomously generates novel themes, angles, and voices for stories, sending them regularly to subscribed journalists and editors through existing channels.
  2. VIREO (Human Opsis, SME from Greece): Recommending images to professionals in the News and Media industry using AI techniques, enhancing the creation of visually compelling articles and improving the reading experience for media consumers.
  3. NLMIE (Kaspar Aps, SME from Denmark): Combining Natural Language Processing with Computer Vision to modernize audiovisual archives.
  4. MBD (Tech Wave Development Srl, SME from Romania): Uniting artists, journalists, and programmers against misinformation by providing ways to visualize the hidden structures of fake information, making it meaningful for both journalists and the general public.
  5. magnet (inknow solutions, lda, SME from Portugal): Offering a tool to support journalists in the early phases of article production by automatically resurfacing relevant content from previous activities.


RESEARCH projects:

  1. CAMOUFLAGE (Politecnico di Torino, Higher Education institution from Italy): Developing diffusion models for extreme image anonymization in social media.
  2. ELMER (University of Surrey, Higher Education institution from the UK): Creating an efficient system for content retrieval capable of handling multi-modal audio, image, text, and video data, particularly for footage longer than 10 seconds.
  3. HoloNeXT (Fundació i2CAT, Research organization from Spain): Developing a novel XR media production tool integrating two volumetric/XR technologies: Neural Radiance Field scene modeling and holographic real-time video volumetric transmission.
  4. CLIP LENS (CENTIC, Research organization from Spain): Enhancing AI-based systems like image classifiers and search engines through generative data augmentation and CLIP.
  5. VolEvol (“Gheorghe Asachi” Technical University of Iasi, Higher Education institution from Romania): Facilitating the rendering of images from volume data using evolutionary algorithms to search for rendering parameters based on quality and diversity-oriented optimization objectives.

Authors: Samuel Almeida, Ellie Shtereva, and Catarina Reis (F6S)

The AI Media Observatory is now fully launched

The AI Media Observatory is a knowledge platform that monitors and curates relevant research on AI in media. Over the last few months, we have been slowly building a knowledge foundation of articles and audiovisual content covering questions regarding the environmental and societal impact of AI, emerging policies and legislation, how to ensure social and ethical AI, and what the upcoming trends and technologies look like.

The content currently featured on the observatory is curated by the consortium and is based on the expertise of more than 30 leading research and industry partners in the field of AI in media. However, all stakeholders are also invited to submit content to the Observatory to ensure its relevance for everyone working at the intersection of media and AI. More information on how to submit and the criteria can be found on the ‘Editorial Board’ page.

As the latest feature, the AI Media Observatory now also includes an expert directory where AI and Media experts can be featured, allowing stakeholders to easily get in touch with relevant experts in the field. The expert directory is open, and all experts are welcome to sign up and have their profiles featured as long as they meet the eligibility criteria.

The Observatory in short

In short, the overarching goal of the Observatory is to support the ongoing efforts of the multidisciplinary community of professionals who are working towards ensuring the responsible use of AI in the media sector. It aims to contribute to the broader discussion and understanding of the development and use of AI in the sector and its impacts on society, the economy, and people. The observatory aims to fulfil this goal by curating relevant content that provides expert perspectives on the potentials and challenges that AI poses for the media sector through its sections ‘Your AI Media Fee’ for written content and ‘Let’s Talk AI and Media’ for audiovisual content. It also provides an easy overview of relevant experts in the field through our directory ‘Find your AI Media Expert.’

Author: Anna Schjøtt Hansen (UvA)