C2PA may end up stripping meme lords of their anonymity


Alright, picture this: You are the ultimate meme lord! You’ve got this epic talent for creating viral masterpieces using third-party images. Or videos. You know the secret sauce – adding those side-splitting, sarcastic, or conversion captions that instantly turn any image or video into internet gold. And guess what? You don’t shy away from a bit of controversy! You’ve got that fearless streak in you, and you’re not afraid to throw some playful jabs, even at the president! You’ve got everyone talking, sharing, and laughing out loud, all thanks to your ingenious memes.

There’s more! You recently unlocked a whole new level of meme wizardry by harnessing the power of generative AI. Now you can create lifelike photos of celebrities and present them in the most hilarious and unexpected ways, like imagining them in the nude! Your creativity knows no bounds, and the internet can’t get enough of it. And what makes it all even more thrilling is the fact that you’re like a digital ninja – anonymous and mysterious. No one knows your true identity, not the make or model of the laptop where you brew your meme magic. How about never ever getting to figure out where your not-so-humble abode is? It’s like you’re an internet ghost, effortlessly blending in the digital shadows.

But hold onto your keyboard, because there’s a twist in this tale. A new player is stepping onto the scene, and they go by the name of C2PA. The C2PA is a joint initiative by Adobe, Arm, Intel, Microsoft, Truepic and the BBC set up to develop technical standards for certifying the source and history (or provenance) of media content. The C2PA was founded in February 2021 as a merger of two previous efforts: the Adobe-led Content Authenticity Initiative (CAI) and the Microsoft- and BBC-led Project Origin.

The problems that the C2PA seeks to address are not new, but they have become more urgent and complex with the emergence of generative AI. Right now, anyone can create, edit, and distribute content with ease, giving rise to questions such as how can we tell what is true and what is not. How can we distinguish between authentic and manipulated content? How can we protect ourselves from misinformation and disinformation that can undermine democracy, health, or security? These questions arise because of the ever-proliferation of fake content that includes:

  • Deepfakes: videos or images that use artificial intelligence to replace the face or voice of a person with another one
  • Cheapfakes: videos or images that use simpler techniques to alter the appearance or context of a person or an event
  • Misinformation: false or inaccurate information that is spread unintentionally or without malice
  • Disinformation: false or inaccurate information that is spread deliberately or with malice

Manipulated content can have serious consequences for individuals, organizations, and society at large. It can erode trust in institutions, undermine the credibility of sources, damage reputations, influence opinions, sway elections, incite violence, and endanger lives. According to a recent report by the World Economic Forum, online misinformation and disinformation are among the top global risks in terms of likelihood and impact.

The C2PA proposes a solution that is based on the concept of provenance. Provenance is the information that describes the origin, creation, modification, and distribution of a piece of content. Provenance can help users verify the authenticity and integrity of content by answering questions such as:

  • Who created the content?
  • When and where was the content created?
  • How was the content created?
  • What changes were made to the content after creation?
  • Who distributed the content?

The C2PA aims to develop technical standards for capturing, storing, and verifying provenance information for different types of media content, such as images, videos, audio, text, and documents. The standards will be open, interoperable, scalable, secure, and privacy-preserving. The standards will also enable users to control how much provenance information they want to share or access.

The C2PA envisions a future where provenance information is embedded in every piece of content that is created or shared online. This will allow users to access provenance information through various interfaces, such as browsers, apps, platforms, or devices. Users will be able to see visual indicators or metadata that show the provenance status of content. Users will also be able to verify the provenance information using cryptographic methods or third-party services.

The C2PA believes that provenance standards can bring many benefits to different stakeholders in the online ecosystem. Some of these benefits include:

  • For creators: Provenance standards can help creators protect their intellectual property rights and attribution. They can also help creators showcase their originality and credibility.
  • For publishers: Provenance standards can help publishers enhance their reputation and trustworthiness. They can also help publishers comply with ethical and legal standards for journalism.
  • For platforms: Provenance standards can help platforms reduce the spread of harmful content and improve their moderation policies. They can also help platforms provide better user experiences and engagement.
  • For consumers: Provenance standards can help consumers make informed decisions about what content to trust and what content to avoid. They can also help consumers exercise their rights to privacy and consent.

The C2PA however isn’t a silver bullet for solving all the problems of online misinformation and disinformation. One particular challenge that C2PA is likely to deal with is the potential eradication of anonymity. How will the C2PA affect the ability of users to create and consume content anonymously? How will the C2PA protect the privacy rights and preferences of users? How will the C2PA balance the need for transparency and accountability with the need for anonymity and confidentiality?

Though anonymity enables cowards to be deceptive, and manipulative online, it is also credited for enabling freedom of expression, creativity, diversity, whistleblowing, and activism. That’s why a technology or standard that is likely to eradicate anonymity must worry any online user. If any content generated or modified by anyone will carry with it tags that can be used to potentially identify the creator or modifier of the content, won’t this allow governments and big organizations to crack down on the dissenting voices?

C2PA’s answer to this dilemma is that it does not aim to eliminate anonymity online, but rather to provide users with more information and control over the provenance of content. The standard claims that it will only allow users to choose how much provenance information they want to share or access. For example, a user who creates content can decide whether to include their identity or not in the provenance metadata. A user who consumes content can decide whether to trust content that has no or limited provenance information or not.

At the end of the day, at stake are two fundamental pillars: the safety of individuals from the spreading menace of falsehoods using their identities and the preservation of our precious human rights like privacy and freedom of speech. What do we truly value more? Are we willing to trade some privacy and freedom to protect ourselves from the potential harm caused by malicious actors manipulating identities and disseminating deceit? Or do we stand firm, holding fast to our rights, even if it means facing the challenges posed by misinformation and its adverse effects?

In this digital era, where information can shape destinies and perceptions, the path we choose will redefine the landscape of communication and the flow of information. It’s not just a matter of technical implementation; it’s a profound ethical and societal choice.

Between Raila and Ruto, who gains more from the violent protests?


Welcome! Login in to your account

Remember me Lost your password?

Lost Password