“Seeing is believing” has been revoked by deep fake technology

A landscape concept art, smooth and dramatic, in a wide angle, depicting 'The dangerous potential of deep fakes.' The scene illustrates a human figure
(Photo: AI-Generated)
A landscape concept art, smooth and dramatic, in a wide angle, depicting 'The dangerous potential of deep fakes.' The scene illustrates a human figure

Hamza Alakaleek

Hamza Alakaleek has graduate degrees in International Political Economy and International Business Law from Yarmouk University and University de Montreal with focus in Internet of Things, Artificial Intelligence and Data Protection.

There is a strong perception that with the advent of technology, the possibilities of cyber-security breaches, hacking of personal information, and online financial frauds would increase dramatically. Further, the introduction of deepfakes has raised major ethical issues as it has the potential to destroy our old deep-rooted belief that “seeing is believing”. Now, it is seemingly impossible to believe even after seeing an act online. This would greatly infringe on an individual’s ability for decision-making and is viewed as highly disturbing.اضافة اعلان

So, what can the technology industry do to survive the numerous challenges, stay relevant, and grow as dramatically as it has been doing for the past many years? Are there many choices besides continuously innovating, being agile, and maintaining a dynamic balance while grappling with critical perceptions, shifting realities, and unexpected disruptions in the global economy.

Eight Years ago brought a sea change in the U.S. and Europe to our understanding of the impact of nation-state disinformation campaigns on major social media platforms. December brought the most definitive assessment yet, with a report sponsored by the Senate Intelligence Committee using data provided under subpoena by the major platforms. Among other things, it showed that between 2015 and 2017, more than 30 million users shared state-sponsored Facebook and Instagram posts with their family and friends, liking, reacting to, and commenting on them along the way.
The introduction of deepfakes has raised major ethical issues as it has the potential to destroy our old deep-rooted belief that “seeing is believing”.
Deepfake content is increasing and fueling the spread of false information. As people become aware that deepfake content may be produced and disseminated, frequently with malicious intentions, they are beginning to doubt the public’s perception of video as legitimate content. Businesses must be ready with verification tools to ensure their legitimate content is not used improperly, for example, when Israel launched a strike at the Ahli Arab Hospital in Gaza City many Western news organizations, including The New York Times, reported the Israeli military’s assertion that the blast was the result of a failed rocket launch by Palestinian Islamic Jihad, an armed group aligned with Hamas, but the Palestinian supports dinged deep to clarify the truth and the used many verification techniques to prove that the IDF is the one who hit the hospital and killed more than 200 innocent civilians relying on evidence like a tweet of a IDF general commander.

Ethical dilemmas
Businesses today are faced with several ethical challenges. Critical decisions have to be made to ensure we are protecting personal freedoms and using data appropriately. Which ethical issues are the most important in 2023?

One of the primary ethical dilemmas in our technologically empowered age revolves around how businesses use personal information. As we browse internet sites, make online purchases, enter our information on websites, engage with different businesses online, and participate in social media, we are constantly providing personal details. Companies often gather information to hyper-personalize our online experiences, but to what extent is that information actually impeding our right to privacy?

Personal information is the new gold, as the saying goes. We have commoditized data because of the value it provides to businesses attempting to reach their consumer base. But when does it go too far? For businesses, it’s extremely valuable to know what kind of products are being searched for and what type of content people are consuming the most. For political figures, it’s important to know what kind of social or legal issues are getting the most attention. These valuable data points are often exploited so that businesses or entities can make money or advance their goals. Facebook in particular has come under fire several times over the years for selling personal data it gathers on its platform.

One thing that became evident during the 2016 and 2020 U.S. presidential elections was the potential of misinformation to gain a wider support base. The effect created polarization that has had wide-reaching effects on global economic and political environments. In contrast to how information was accessed prior to the internet, we are constantly flooded with real-time events and news as it breaks. Celebrities and political figures can disseminate opinions on social media without fact-checking, which is then aggregated and further spread despite its accuracy—or inaccuracy. Information no longer undergoes the strenuous validation process that we formerly used to publish newspapers and books.
Ultimately, we need to create a culture of responsibility within technology. If the information technology workforce and industry giants believe they are responsible for the safe and ethical usage of technology, then we will see more governance and fair use of data.
Similarly, we used to believe that video told a story that was undeniably rooted in truth. But deepfake technology now allows such a sophisticated manipulation of digital imagery that people appear to be saying and doing things that never happened. The potential for privacy invasion and misuse of identity is very high with the use of this technology.

Taking responsibility
Most companies operate with a hybrid stack, comprised of a blend of third-party and owned technology. As a result, there is often some confusion about where responsibility lies when it comes to governance, the use of big data, cybersecurity concerns, and managing personally identifiable information or PII. Whose responsibility is it really to ensure data is protected? If you engage a third party for software that processes payments, do you bear any responsibility if credit card details are breached? The fact is that it’s everyone’s job. Businesses need to adopt a perspective where all collective parties share responsibility.

Unlike business ethics, ethical technology is about ensuring there is a moral relationship that exists between technology and users. Businesses that engage in ethical technology have a firm moral sense of employee rights and customer protection. Data is valuable, but the employees and customers who power your business are undoubtedly your greatest asset.

Data is undoubtedly something of value for businesses. It allows companies to target their marketing strategies and refine product offerings, but it can also be an invasive use of privacy bringing many ethical considerations to the forefront. Data protection measures and compliance procedures can help ensure that data isn’t leaked or used inappropriately. Ultimately, we need to create a culture of responsibility within technology. If the information technology workforce and industry giants believe they are responsible for the safe and ethical usage of technology, then we will see more governance and fair use of data.

Since the way to the solution isn’t a straight line, it is obvious that there isn’t a single solution. However, as we move into 2023 and beyond, information technology will become increasingly important to how we work, play, and live. The capacity of tech organizations to prosper in the upcoming years will undoubtedly be impacted by how they continue to support clients with their digital transformations and how they manage and accept new technology.


Dr. Hamza Alakaleek is a Corporate lawyer and tax consultant with post-graduate degrees in international political economy, international business law, and law and technology with a focus on internet of things, artificial intelligence and data protection.


Disclaimer: 
Views expressed by writers in this section are their own and do not necessarily reflect Jordan News' point of view.



Read more Opinion and Analysis
Jordan News