Can artificial intelligence be dangerous?


On May 17, Sam Altman, the CEO of OpenAI, the company behind ChatGPT, addressed the US Senate committee, advocating for the regulation of artificial intelligence. Altman emphasized the transformative potential of AI, likening it to the invention of the printing press, but also stressed the importance of addressing its potential risks. A US senator echoed these concerns, likening AI’s dangers to those of an atomic bomb.

The launch of ChatGPT on November 30, 2022, marked a significant milestone in technology. This AI platform has quickly gained traction across various sectors, including technology, health, education, commerce, business, and banking, amassing a user base of 100 million within a short span of time.

ChatGPT is a sophisticated AI tool and language model capable of human-like communication and answering questions. In addition to processing text, it can now handle images and videos and even design websites based on simple sketches. Trained on a vast dataset comprising three billion words, ChatGPT continuously learns and improves through interactions with users, performing tasks ranging from composing emails to conducting audits.

The widespread adoption of ChatGPT by companies, including integration with Microsoft’s Bing search engine and incorporation into Snapchat’s app, has further propelled its popularity and influence.

The proliferation of ChatGPT sparked widespread discussions on social media platforms in Pakistan, with users engaging in conversations and sharing screenshots of their interactions with the AI. Many found the experience unbelievable and surprising.

There were speculations among some individuals, including a friend of mine, who believed there was a human operator behind ChatGPT. However, after thorough discussion, I was able to convince him that it was merely an automated software.

While technological innovations are indicative of societal progress, they necessitate regulation to safeguard public welfare. This is especially crucial when considering the potential for AI to disrupt the balance between humans and machines, potentially leading to undesirable power dynamics.

Analogously, the relationship between humans and mobile phones illustrates the need for regulations to govern their usage, as they have become indispensable in our daily lives. Similarly, without timely regulations for artificial intelligence, there is a risk of being overwhelmed by its influence.

Although artificial intelligence offers numerous benefits, its unregulated use poses risks across various domains, including politics, society, economy, and military. In politics, chatbots can be employed to assess public sentiment towards different candidates, yet their unregulated use can lead to manipulation and misinformation.

Artificial intelligence (AI) plays a significant role in political communication, offering benefits such as assisting political leaders in running informed campaigns, managing websites and social media accounts, and interacting with constituents. Generative AI can even analyze public sentiment and predict election outcomes in advance.

However, regulated AI systems can also be exploited to impersonate political figures, manipulate public opinion, and disseminate misinformation. Given AI’s potential to influence political discourse, it’s crucial for policymakers to consider ethical implications and address associated risks. Despite being data-driven, AI systems can produce false or distorted content, as demonstrated by recent events like the manipulated video of Ukrainian President Volodymyr Zelensky.

Moreover, AI and chatbots can be used to spread fake news and manipulate social media algorithms, posing serious threats to public safety. Despite these concerns, AI has the potential to boost productivity and efficiency in various economic sectors, with generative AI projected to increase the global GDP by 7.0% over the next decade.

However, the rise of AI also raises apprehensions about job displacement, as intelligent AI models like Google Cabard and ChatGPT possess capabilities surpassing those of many humans. While AI augments human capabilities in several domains, there remains a persistent risk of automation replacing human workers.

The advancement of artificial intelligence (AI) has led to transformations across various sectors, including transportation and education. While AI-driven technologies like automated cars may render certain jobs redundant, they also create new technical and non-technical job opportunities. Adaptability to these changes is crucial for individuals to remain competitive in the evolving job market.

Similarly, AI, exemplified by ChatGPT, finds applications in diverse social sectors, notably education. Students leverage AI tools to facilitate tasks such as assignment creation and subject summarization, alleviating their workload but potentially burdening educators. The challenge lies in ensuring the integrity of AI-generated content, as reliance solely on AI may compromise educational standards and meritocracy. Thus, regulations are necessary to uphold honesty and fairness in the educational realm.

Moreover, AI holds promise in military applications, offering significant advantages in software development and decision-making processes. Military agencies recognize AI’s potential to enhance defense strategies, analyze sensor data, and optimize emerging technologies like drones, missiles, and tanks. Despite its civilian origins, AI’s integration into the military sector underscores its multifaceted utility across various domains.

The utilization of generative AI in lethal automated weapons systems raises profound concerns regarding accountability, responsibility, and transparency. These systems have the potential to autonomously target individuals once activated, raising questions about who bears responsibility for their actions. Consequently, stringent regulation is essential to mitigate the risk, particularly as AI becomes increasingly intelligent and capable of independent decision-making, potentially circumventing human oversight.

While Italy temporarily banned ChatGPT, such prohibitions are not widespread, with many countries hesitant to enact laws regulating generative AI for fear of hindering technological advancement. However, in an era marked by rapid technological development, discussions of banning technology are deemed impractical and counterproductive, particularly for technologically advanced nations.

Instead, emphasis should be placed on implementing robust regulations and laws, as advocated by the CEO of OpenAI and other prominent figures in Silicon Valley. Despite some progress, as indicated by the passage of approximately 37 bills globally addressing limited aspects of AI regulation, comprehensive frameworks are lacking at the international level.

Amidst escalating global competition in AI, the momentum for regulation appears to be waning, posing potential risks for the future. Only through the establishment of effective laws and regulations can the positive and beneficial utilization of AI be ensured, safeguarding against its potential misuse and preventing catastrophic consequences akin to those of the atomic bomb.

Leave a Reply

Your email address will not be published. Required fields are marked *