Spotify to Lay Off 1,500 Employees

Spotify says it is planning to lay off 17% of its global workforce, amounting to around 1,500 employees, following layoffs earlier this year of 600 people in January and an additional 200 in June.

The music streaming giant is continuing its effort to cut costs and work toward becoming profitable, said Spotify CEO Daniel Ek in a prepared statement.

“By most metrics, we were more productive but less efficient,” he said. “We need to be both.”

The layoffs come following a rare quarterly net profit of about $70.3 million in October. The company has never seen a full year net profit.

“I realize that for many, a reduction of this size will feel surprisingly large given the recent positive earnings report and our performance,” Ek said. “We debated making smaller reductions throughout 2024 and 2025. Yet, considering the gap between our financial goal … and our current operational costs, I decided that a substantial action to right size our costs was the best option to accomplish our objectives.”

With the new layoffs, the company now expects to see a fourth quarter loss between $100 million to $117 million after previously anticipating a $40 million profit.

A majority of the charges will go toward severance for laid off employees, who will get about five months’ pay, vacation pay and health care coverage for the severance period.

Spotify did not clearly state when the layoffs would become financially beneficial but said that they would “generate meaningful operating efficiencies going forward.”

Spotify is following many companies in the tech industry trying to cut costs after growth in the industry slowed following a surge during the COVID pandemic.

Tech giants including Meta, Microsoft, Amazon and Google parent company, Alphabet, all have plans to cut 10,000 or more people this year.

Spotify began informing affected employees on Monday.

Some information in this report came from Reuters, The Associated Press and Agence France-Presse.

your ad here

Solar Grid Brings Light, Progress to Rural Nigerian Community

In early November, Nigeria launched its first interconnected solar hybrid mini grid in Nasarawa State to make electricity more reliable, renewable and accessible. Gibson Emeka has this story from Abuja, Nigeria, Grace Oyenubi narrates.

your ad here

Breaches by Iran-Affiliated Hackers Span US States, Federal Agencies Say

A small western Pennsylvania water authority was just one of many organizations breached in the United States by Iran-affiliated hackers who targeted a specific industrial control device because it is Israeli-made, U.S. and Israeli authorities say.

“The victims span multiple U.S. states,” the FBI, the Environmental Protection Agency, the Cybersecurity and Infrastructure Security Agency, known as CISA, as well as Israel’s National Cyber Directorate said in an advisory emailed to The Associated Press late Friday.

They did not say how many organizations were hacked or otherwise describe them.

Matthew Mottes, the chairman of the Municipal Water Authority of Aliquippa, which discovered it had been hacked on Nov. 25, said Thursday that federal officials had told him the same group also breached four other utilities and an aquarium.

Cybersecurity experts say that while there is no evidence of Iranian involvement in the Oct. 7 attack into Israel by Hamas that triggered the war in Gaza, they expected state-backed Iranian hackers and pro-Palestinian hacktivists to step up cyberattacks on Israeli and its allies in its aftermath. And that has happened.

The multiagency advisory explained what CISA had not when it confirmed the Pennsylvania hack Wednesday — that other industries outside water and water-treatment facilities use the same equipment — Vision Series programmable logic controllers made by Unitronics — and were also potentially vulnerable.

Those industries include “energy, food and beverage manufacturing and healthcare,” the advisory says. The devices regulate processes including pressure, temperature and fluid flow.

The Aliquippa hack promoted workers to temporarily halt pumping in a remote station that regulates water pressure for two nearby towns, leading crews to switch to manual operation. The hackers left a digital calling card on the compromised device saying all Israeli-made equipment is “a legal target.”

The multiagency advisory said it was not known if the hackers had tried to penetrate deeper into breached networks.

The advisory says the hackers, who call themselves “Cyber Av3ngers,” are affiliated with Iran’s Islamic Revolutionary Guards Corps, which the U.S. designated as a foreign terrorist organization in 2019.

The group targeted the Unitronics devices at least since Nov. 22, it said.

An online search Saturday with the Shodan service identified more than 200 such internet-connected devices in the U.S. and more than 1,700 globally.

The advisory notes that Unitronics devices ship with a default password, a practice experts discourage as it makes them more vulnerable to hacking. Best practices call for devices to require a unique password to be created out of the box. It says the hackers likely accessed affected devices by “exploiting cybersecurity weaknesses, including poor password security and exposure to the internet.”

In response to the Aliquippa hack, three Pennsylvania congressmen asked the U.S. Justice Department in a letter to investigate. Americans must know their drinking water and other basic infrastructure is safe from “nation-state adversaries and terrorist organizations,” U.S. Sens. John Fetterman and Bob Casey and U.S. Rep. Chris Deluzio said.

Cyber Av3ngers claimed in an Oct. 30 social media post to have hacked 10 water treatment stations in Israel, though it is not clear if they shut down any equipment.

Unitronics has not responded to the AP queries about the hacks.

The attack came less than a month after a federal appeals court decision prompted the EPA to rescind a rule that would have obliged U.S public water systems to include cybersecurity testing in their regular federally mandated audits. The rollback was triggered by a federal appeals court decision in a case brought by Missouri, Arkansas and Iowa, and joined by a water utility trade group.

The Biden administration has been trying to shore up cybersecurity of critical infrastructure — more than 80% of which is privately owned — and has imposed regulations on sectors including electric utilities, gas pipelines and nuclear facilities. But many experts complain that too many vital industries are permitted to self-regulate.

your ad here

Rules Would Bar EV Tax Credits if Batteries, Minerals Linked to China

The U.S. proposed new guidelines Friday spelling out which electric vehicles will be eligible for tax credits, ruling out those that contain batteries or minerals sourced from China and other nations that have fallen out of favor with the U.S.

The restrictions dictate which clean energy vehicles will qualify for a subsidy of up to $7,500 under President Joe Biden’s Inflation Reduction Act, a federal law promoting sustainable, domestic energy production.

Only about 20 out of the more than 100 electric vehicles on the U.S. market qualify for a tax credit as it is. That number may be further reduced when this regulation goes into effect.

If a clean energy battery went through an assembly line owned by any “foreign entity of concern,” the car it will go into would be immediately disqualified from earning its owner any tax breaks from the U.S. government, starting in 2024.

The new rules target firms incorporated or headquartered in China, Russia, North Korea and Iran, among others, as well as companies where 25% or more of the equity interest or board seats are controlled by those countries.

From 2025 onward, electric vehicles made with critical minerals, such as lithium, nickel and cobalt, mined or processed by any “foreign entity of concern” will also be ineligible for subsidies.

The rules will be open to public feedback from automotive leaders for several weeks and are subject to change depending on industry recommendations.

Some information for this report came from Agence France-Presse. 

your ad here

US Lawmakers Say Chinese Disinformation Operations a Growing Threat

US lawmakers warned Thursday that the Chinese Communist Party’s efforts to spread disinformation will only increase ahead of the 2024 elections in the United States. As VOA’s Congressional Correspondent Katherine Gypson reports, social media giant TikTok is lawmakers’ top concern.
Camera: Saqib Ui Islam

your ad here

VOA Exclusive: US, S Korea, Japan to Sign Pact to Counter Disinformation  

The United States plans to sign a memorandum of understanding to cooperate with South Korea and Japan in the fight against false propaganda and disinformation.

It will be the first such agreement that Washington signs with its Asian allies, and it comes as U.S. officials and lawmakers accuse the People’s Republic of China of conducting “deceptive online campaigns” targeting the United States and other countries. Chinese officials have rejected the accusation.

Liz Allen, the U.S. undersecretary of state for public diplomacy and public affairs, is traveling to Asia this week. Allen will be sealing the agreement with South Korea and Japan on countering disinformation, according to U.S. and diplomatic sources.

U.S. President Joe Biden, South Korean President Yoon Suk Yeol and Japanese Prime Minister Fumio Kishida have agreed to find ways to coordinate efforts to counter disinformation, after the three leaders held talks during their first trilateral summit at Camp David in August.

“President Yoon mentioned the threat from false propaganda and disinformation in his address to the joint session of U.S. Congress in April. In this regard, we are now discussing the possible follow-up measures with the U.S.,” an official from the South Korean Embassy told Voice of America on Thursday.

In a statement on Thursday, House Foreign Affairs Committee Chairman Michael McCaul condemned the “increasingly deceptive online campaigns targeting the U.S. and other countries” by the Chinese Communist Party.

“The CCP has made clear it will use every tactic to spread its malign intent,” the Republican congressman said.

The South Korean government has identified 38 suspected fake Korean-language news websites that it believes are operated by Chinese companies. For example, in November, South Korea’s National Intelligence Service said two Chinese public relations companies, Haimai and Haixun, were allegedly creating such websites, according to Seoul-based Yonhap News Agency.

The State Department said Allen, while in Tokyo, will hold bilateral discussions with Japanese Ministry of Foreign Affairs officials that include a focus on countering malign foreign influence.

In a report issued in September, the State Department’s Global Engagement Center accused the Chinese government of using a combination of tactics in a bid to create a world in which Beijing, either explicitly or implicitly, controls the flow of critical information. The U.S. has warned that China is pouring billions of dollars into efforts to reshape the global information environment and, eventually, bend the will of multiple nations to Beijing’s advantage.

The Chinese Ministry of Foreign Affairs has pushed back, saying the report by U.S. State Department’s Global Engagement Center “misrepresents facts and truth.” A spokesperson from the Chinese Foreign Ministry called GEC the command center of “perception warfare.”

James Rubin, special envoy for the State Department’s Global Engagement Center, has said that Washington is working with allies to detect and counter misinformation and disinformation around the world.

In May, the U.S. signed a memorandum of understanding with North Macedonia, and in September, another with Bulgaria, both aimed at enhancing cooperation in countering foreign information manipulation.

your ad here

Vietnam’s Rare Earth Sector on the Rise

Vietnam, with the world’s second-largest reserves of the rare earths used in such modern devices as electric vehicle batteries and smart phone screens, is intensifying mining of the critical minerals. The industry, though, faces high processing costs, environmental concerns, and the takedown of industry leaders for illegal mining and mineral sales.

Vietnam’s rare earth resources are second only to those of China, which has held a tight monopoly since the 1980s. With Chinese relations with the West becoming more volatile, many countries are looking for other sources for the elements.

“China produces about 60% of the world’s rare earths but what they process is over 90%,” Louis O’Connor, CEO of Strategic Metals Invest, an Irish investment firm, told VOA.

“It was not a good idea to allow one country to dominate critical raw materials that are critical to all nations’ economic prosperity and increasingly military capability,” he said.

O’Connor added that while China has the world’s majority of raw materials, its dominance over technical expertise in the complex and costly process of rare earth refining is even greater. China has 39 metallurgy universities and approximately 200 metallurgists graduate weekly in the country, he said.

“The ability to go from having the potential to end product — that’s the most challenging, complicated, and expensive part,” O’Connor said. “For Vietnam, even if they have the deposits, what they don’t have is the human capital, or the engineering expertise.”

Vietnam increased rare earth mining tenfold with its output hitting 4,300 tons last year, compared to 400 tons in 2021.  according to the U.S. Geological Survey. Vietnam said in July it plans to process 2 million tons of rare earth ores by 2030 and produce 60,000 tons of rare earth oxides annually starting in 2030. This year, China’s mining quota is set at 240,000 tons to meet the demand for the electric vehicle industry, according to Chinese government data.

The United States and other countries are interested in Vietnam increasing its production of rare earths.

“The U.S. wants Vietnam to become a more important supplier and perhaps replace China, if possible, because of the risk that the U.S. may face in relying on rare earth supplies from China,” Le Hong Hiep, senior fellow at the ISEAS-Yusof Ishak Institute in Singapore told VOA.

“Not only the U.S., but also other partners like Korea, Japan, and Australia also are working with Vietnam to develop the rare earth industry,” he said.

South Korean President Yoon Suk Yeol signed a memorandum of understanding in Vietnam in June to establish a joint supply chain center for rare earth minerals.

“We reached an agreement that there is more potential to develop rare earths together, as they are abundant in Vietnam,” Yoon said in a June 23 statement with Vietnam’s president Vo Van Thuong.

The United States signed such a memorandum on cooperation in the rare earths sector during President Joe Biden’s visit to Hanoi on September 9.

“We see Vietnam as a potential critical nexus in global supply chains when it comes to critical minerals and rare earth elements,” Marc Knapper, U.S. ambassador to Vietnam, said on September 13 during a digital press briefing. “We certainly want to work together to ensure that Vietnam is able to take advantage of its rich resources in a way that’s also sustainable.”

Scandal

However, a handful of Vietnam’s key rare earth enterprises have become entangled in scandal. On October 20, police arrested six individuals for mining and tax violations.

Police arrested Doan Van Huan, chairman of the Hanoi-based Thai Duong Group that operates a mine in Yen Bai province, and its chief accountant, Nguyen Van Chinh, for violating regulations on the exploration and exploitation of natural resources and accounting violations, the Public Security Ministry said. The two were accused of making $25.5 million from the illegal sale of rare earth ore and iron ore. Police raided 21 excavation and trading sites in Yen Bai province and three other locations. Authorities seized an estimated 13,700 tons of rare earth and more than 1,400 tons of iron ores, according to local publication VnExpress.

Although government statements did not state what made Thai Duong’s rare earth sales illegal, a source told Reuters raw Yen Bai mine ore had been exported to China to avoid high domestic refining costs, in violation of Vietnamese rules.

The chairman, Luu Anh Tuan, and accountant, Nguyen Thi Hien, of the country’s primary rare earth refining company, Vietnam Rare Earth JSC, were also arrested for violating accounting regulations in trading rare earth with Thai Duong Group. Dang Tran Chi, director of Hop Thanh Phat, and his accountant Pham Thi Ha were arrested on the same charge.

Looking at corruption in Vietnam’s rare earth industry will be “top of the list” for future investors, O’Connor said.

“Corruption levels would have to be looked at,” he said. “If you’re buying a metal that’s going to need to perform in a jet engine, for example or a rocket … they have to be sure of the purity levels. The chain of custody of these, it’s more important really than gold.”

Vietnam committed to industry

Hanoi is committed to developing the rare earths industry even though economic gains are limited by environmental and production costs, Hiep told VOA.

“Vietnam is now interested in promoting this industry mainly because of the strategic significance,” Hiep told VOA. “If you can grow this industry and become a reliable supplier of rare earth products for the U.S. and its allies, Vietnam’s strategic position will be enhanced greatly.”

“Whether that will be successful, we have to wait and see,” he added.

There are also environmental concerns for the growing industry, particularly as a crackdown on Vietnam’s environmental organizations and civil society leaves little room for public speech.

“The biggest challenge is going to be how do you handle the waste process from the mining,” said Courtney Weatherby, deputy director of the Southeast Asia Program at Washington’s Stimson Center told VOA.

“Ensuring that development happens in a sustainable way takes a lot of different actors,” she said.

But Duy Hoang, executive director of unsanctioned political party Viet Tan, said the room for outside actors to express concern over environmental and labor practices is narrowing.

“What we’re seeing is sort of a shrinking space for civil society to speak out and a number of the leading environmental activists are now in jail. We don’t have their voices which are very needed and I think there may be self-censorship going on by other activists,” he said. “There has to be accountability.”

your ad here

White House Hopes to Lead Global Charge in ‘Promise, Peril’ of Emerging Tech Like AI

American leadership is essential in establishing norms and laws to “determine how we both glean the promise and manage the peril” of emerging technologies like artificial intelligence and digital economic and social platforms used to connect billions of people around the world, a White House adviser told VOA.

The Biden administration has rolled out a number of initiatives on the topic — most recently, an executive order that aims to set new AI safety and security standards. That order relies on cooperation from private developers and other countries, “because the attackers are in one set of countries, the infrastructure is in another and the victims are global,” said Anne Neuberger, deputy national security adviser for cyber and emerging technology at the National Security Council.

Neuberger sat down with VOA White House Correspondent Anita Powell to explain these complex, compelling technologies and how she thinks they have exposed the worst but also the best in humanity.

The following has been edited for length and clarity.

VOA: Thank you so much for sitting down with VOA today. Can you walk us through the concrete outcomes of the recent meeting between President Joe Biden and Chinese President Xi Jinping in the areas you cover — cybersecurity, AI and the digital economy?

Anne Neuberger, White House deputy national security adviser: Of course, strategic technologies are very important to both of our countries’ growth and national security — and we’re global players on a global stage. The most important part of the discussion was two leaders coming together to say: While we are in competition, we’re committed to working together on areas where we can collaborate – areas like climate change, like discussions of what are the rules for artificial intelligence.

VOA: Would you assess that the meeting made any progress, especially on AI regulation?

Neuberger: Certainly very good discussions related to an agreement for the countries to sit down and establish a working group on AI [about] appropriate guardrails and guidelines in this area.

VOA: I’m going to stick with AI and the administration’s recent moves, like the AI Bill of Rights and also the attempt to set some norms at the recent London summit on AI. Why does the administration think U.S. leadership matters so much here?

Neuberger: For two reasons. First, the United States is a committed democracy and AI is a major technology that brings both promise and peril. It is up to us to determine how we both glean the promise and manage the peril. President Biden has made that clear in his game-changing executive order that, as a country, we must manage the perils in order to glean the promise.

VOA: Speaking of the perils of AI, what is the administration doing to prevent the malicious use of generative AI in both conflicts and contests? I’m talking about conflicts like Israel and Ukraine, but also contests like the upcoming elections in Congo, in Taiwan and here in the United States.

Neuberger: We’ve seen new AI models that generate very realistic videos, very realistic images. In terms of generative AI related to elections, I want to lift up one of the voluntary commitments that the president negotiated, which was around watermarking: having a visible and potentially invisible mark on an AI-generated image or video that notes that this is AI-generated, to alert a viewer. An invisible mark could be used so that, even if there are attempts to remove this mark, the platforms themselves can still be able to portray that message and help educate individuals. This is still an area of evolving technology. It’s getting better and better. But companies made commitments to start marking content that they generate. And I know a number of social media platforms are also making commitments to ensure that they display messages to help consumers who see such content know that it is generated by artificial intelligence.

VOA: Moving on to cybersecurity and malign actors like North Korea and Russia, what is the administration doing to curb their work in this area?

Neuberger: We see North Korea really using cyberattacks as a way to get money because they’re such a heavily sanctioned regime. So North Korea moved from targeting banks to targeting … cryptocurrency infrastructure around the world. And the White House has had a focused effort to bring together all elements we have to fight that with Treasury Department designations.

There’ll be further designations coming up for the cryptocurrency mixers that launder funds stolen from those cryptocurrency infrastructures. We also have been working with the industry to press them to improve the cybersecurity of their systems as well as law enforcement. U.S. law enforcement has been cooperating with partners around the world to take down that server infrastructure and to arrest the individuals who are responsible for some of this activity.

VOA: Tell us a little bit about the counter-ransomware initiative you’re working on.

Neuberger: Absolutely. Essentially, criminal groups, many of which are based in Russia with infrastructure operating from around the world, are locking systems … in order to request that the system owners pay ransom. In the United States alone in the last two years, $2.3 billion was paid in ransom. It’s a fundamentally transnational fight. … What we’ve done is assemble 48 countries, Interpol [and] the European Union to take this on together, because we know that the attackers are in one set of countries, the infrastructure is in another, and the victims are global. As the White House built this initiative, we ensure that the leadership is diverse.

So, for example, the leaders of the effort to build capacity around the world are Nigeria and Germany — intentionally, a country from Africa and a country from Europe, because their needs are different. And we wanted to ensure that as we’re helping countries build the capacity to fight this, we’re sensitive to the different needs of a country like Nigeria, like Rwanda, like South Africa, like Indonesia. Similarly, there’s an effort focused on exercising information, sharing information together.

You asked about the key deliverables from this most recent meeting. I’ll note three big ones. First, we launched a website and a system where countries can collaborate when they’re fighting a ransomware attack, where they can ask for help [or] learn from others who fought a similar attack. Second, we made the first ever joint policy statement — a big deal — 48 countries committing that countries themselves will not pay ransoms, because we know this is a financially driven problem. And third, the United States committed that we would be sharing bad wallets [that] criminals are using to move money around the world so other countries can help stop that money as it moves as well. So that’s an example of three of the many commitments that came out of the recent meeting.

VOA: Let’s talk about the Global South, which has pioneered development of really interesting digital economic technologies like Kenya’s M-PESA, which was rolled out in, like, 2007. Now the U.S. has Venmo, which is modeled on that. How is the U.S. learning from the developing world in the development of these projects and also the perils of these products?

Neuberger: M-PESA is a fantastic example of the promise of digital tech. Essentially, Kenya took the fact that they had a telecom infrastructure, and built their banking infrastructure riding on that, so they leapt ahead to enable people across the country to do transactions online. When you look at Ukraine in the context of Russia’s invasion of Ukraine, Ukraine quickly moved their government online, really building on lessons learned from Estonia, to enable Ukrainians — many who are in Poland and Hungary — to continue to engage with their government in a digital way.

The U.S. Agency for International Development is tremendously proud of that Ukrainian project and is using it as a model as we look to other countries around the world. So we’re learning a lot from the creativity and innovation; what we want to bring to that is American development, skill and aid, and also plugging in American tech companies who can accelerate the rollout of these projects in countries around the world, because we still believe in the promise of digital. But you mentioned the peril, and that’s where cybersecurity comes in.

VOA: This lines us up perfectly for my final question, about the promise and the peril. In the digital world, people can hide behind anonymity and say and do awful things using tools that were meant to improve the world. How do you keep your faith in humanity?

Neuberger: It’s a tremendously important question. It’s one that’s personally important to me. My great-grandparents lost their lives in Nazi death camps. And those members of my family who survived — some survived through the horrors of the camp, some managed to hide out under false identities. And I often think that the promise of digital has also made our identities very evident. Sometimes when I’m just browsing Amazon online, and it recommends a set of books, I think to myself: I wonder how I’d hide if what happened to my grandparents came for me. So as a result, I think that even as we engage with these technologies, we have to ensure that vulnerable populations are protected.

So, the president’s working with AI companies to say companies have an obligation to protect vulnerable populations online, to ensure that we’re using AI to detect where there’s bullying online, where there’s hate speech that goes against common practices that needs to be addressed; where there are AI-generated images related to children or women or other vulnerable populations, that we use AI to find them and remove them; and certainly use law enforcement and the power of law enforcement partnerships around the world to deter that as well. Freedom of speech is a part of free societies. Freedom from harm needs to be a fight we take on together.

VOA: Thank you so much for speaking to our audience.

Neuberger: Thank you.

your ad here

Ukraine Businesses Pivot to New Military Technology Production

Since the start of Russia’s full-scale invasion of Ukraine, drone production in the country has surged. Ukrainian businesses have shifted from manufacturing products for peacetime to producing equipment for wartime. From Kyiv, Myroslava Gongadze explains how Ukrainian ingenuity is altering the course of the war. Camera: Eugene Shynkar.

your ad here

US Imposes Sanctions on Cryptocurrency Mixer Sinbad Over Alleged North Korea Links

The United States on Wednesday imposed sanctions on a virtual currency mixer the Treasury Department said has processed millions of dollars worth of cryptocurrency from major heists carried out by North Korea-linked hackers.

The U.S. Treasury Department in a statement said virtual currency mixer Sinbad, hit with sanctions on Wednesday, processed millions of dollars worth of virtual currency from heists carried out by the North Korea-linked Lazarus Group, including the Axie Infinity and Horizon Bride heists of hundreds of millions of dollars.

Lazarus, which has been sanctioned by the U.S., has been accused of carrying out some of the largest virtual currency heists to date. In March 2022, for example, they allegedly stole about $620 million in virtual currency from a blockchain project linked to the online game Axie Infinity.

“Mixing services that enable criminal actors, such as the Lazarus Group, to launder stolen assets will face serious consequences,” Deputy Treasury Secretary Wally Adeyemo said in the statement on Wednesday.

“The Treasury Department and its U.S. government partners stand ready to deploy all tools at their disposal to prevent virtual currency mixers, like Sinbad, from facilitating illicit activities.”

A virtual currency mixer is a software tool that pools and scrambles cryptocurrencies from thousands of addresses.

Sinbad is believed by some experts in the industry to be a successor to the Blender mixer, which the U.S. hit with sanctions last year over accusations it was being used by North Korea.

The Treasury said Sinbad is also used by cybercriminals to obscure transactions linked to activities such as sanctions evasion, drug trafficking and the purchase of child sexual abuse materials, among other malign activities. 

Wednesday’s action freezes any U.S. assets of Sinbad and generally bars Americans from dealing with it. Those that engage in certain transactions with the mixer also risk being hit with sanctions. 

your ad here

Is AI About to Steal Your Job?

Almost all U.S. jobs, from truck driver to childcare provider to software developer, include skills that can be done, or at least supplemented, by generative artificial intelligence (GenAI), according to a recent report.

GenAI is artificial intelligence that can generate high-quality content based on the input data used to train it.

“AI is likely to touch every part of every job to some degree,” says Cory Stahle, an economist with Indeed.com, which released the report.

The report finds that almost one in five jobs (19.7%) — like IT operations, mathematics and information design — faces the highest risk of being affected by AI because at least 80% of the job skills those positions require can be done reasonably well by GenAI.

But that doesn’t mean that those jobs will eventually be lost to robots.

“It’s important to recognize that, in general, these technologies don’t affect entire occupations. It actually is very rare that a robot will show up, sit in somebody’s seat to do everything that someone does at their job,” says Michael Chui of the McKinsey Global Institute (MGI), who researches the impact of technology and innovation on business, the economy and society.

Indeed.com researchers analyzed more than 55 million job postings and found that GenAI can perform 50% to almost 80% of the skills required in 45.7% of those job listings. In 34.6% of jobs listed, GenAI can handle less than 50% of the skills.

Jobs that require manual skills or a personal touch, such as nursing and veterinary care, are the least likely to be hard hit by AI, the report says.

In the past, technological advances have mostly affected manual labor. However, GenAI is expected to have the most effect on so-called knowledge workers, generally defined as people who create knowledge or think for a living.

But, for now, AI does not appear poised to steal anyone’s job.

“There are very few jobs that AI can do completely. Even in jobs where AI can do many of the skills, there are still aspects of those jobs that AI cannot do,” Stahle says.

Rather than replace workers, researchers expect GenAI to enhance the work people already do, making them more efficient.

“This is something that, in many ways, we believe is going to unlock human potential and productivity for many workers across many different sectors of the economy,” Stahle says.

“There are a number of things that can happen,” Chui adds. “One is, we simply do more of something we were already doing, and so imagine if you’re a university professor or a teacher, and the grading can be done by machine rather than you. You can take those hours and, instead of grading, you can actually start tutoring your students, spending more time with your students, improving their performance, helping them learn.”

American workers need to begin using the new technology if they hope to remain competitive, according to Chui.

“Workers who are best able to use these technologies will be the most competitive workers in the workforce,” he says. “It was true before, but it’s more true than ever, that we’re all going to have to be lifetime learners.”

A survey developed by Chui finds that almost 80% of workers have experimented with AI tools.

“One of the great powers of these generative AI tools, so far, is they’ve been designed in such a way to make it easy for really anybody to use these types of tools,” Stahle says. “I really believe that people should be looking to embrace these tools and find ways to incorporate them into the work that they’re already interested in doing.”

Ultimately, could one of the unexpected benefits of AI be more efficient employees who work less?

“In general, Americans work a lot,” Chui says. “Maybe we don’t have to work so long. Maybe we have a four-day work week … and so you could give that time back to the worker.”

your ad here

US Envoy Focuses on Cyberscams During Cambodia Visit 

Cindy Dyer, the U.S. ambassador-at-large for monitoring and combating trafficking, is planning to push Cambodia’s new government to ramp up its efforts to crack down on cyberscam operations that trap many trafficking victims in slavelike conditions.

A recently completed visit to Phnom Penh by Dyer “will serve as an opportunity for information sharing and coordination on anti-trafficking efforts,” the State Department said last week in a release.

Dyer met with a range of officials “with the objective of building a relationship with the new government for future coordination and advocating for progress in the most critical areas, including increased investigations and prosecutions of cyberscam operations,” said the November 15 release.

Cambodia’s role as host of cybercriminals has been in an international spotlight. The U.N. High Commissioner for Human Rights (UNHCHR) released a report this summer estimating that the industry has victimized 100,000 people in Cambodia.

Lured by promise of jobs

Operators of these scamming networks recruit unwitting workers from across Asia, often with the promise of well-paying tech jobs, and then force them to attempt to scam victims online while living in slavelike conditions, according to the report.

Countries including Indonesia, Taiwan and China have urged countries like Cambodia and Laos to crack down on the industry, while warning their own citizens of the dangers in traveling to these countries, according to the UNHCHR report.

The U.S. State Department’s annual report on global human trafficking, released in June, placed Cambodia in Tier 3, meaning the government has made insufficient efforts to address human trafficking and does not meet the minimum standards.

During her two-day visit to Cambodia that began November 15, Dyer met with officials from the ministries of justice, labor and social affairs, as well as representatives of the National Police and the National Committee for Counter Trafficking (NCCT) within the Ministry of Interior, according to an email from the U.S. Embassy in Phnom Penh. Dyer also held discussions with civil society groups working on combating human trafficking.

The discussions focused “on Cambodia’s efforts to protect trafficking victims, including providing protection assistance services for victims of trafficking and vulnerable migrants, capacity building for service providers and government officials to improve victim identification and referral, and addressing emerging trends in forced criminality,” the State Department release said.

More training urged

Am Sam Ath, operations director at the Cambodian rights group Licadho, told VOA Khmer that Dyer’s visit highlighted the need for Cambodia to tackle human trafficking and online scams.

“We see that the United States … ranks Cambodia third in the blacklist of human trafficking. It also has a lot of impact on our country, and if Cambodia does not make an effort further in the prevention of human trafficking or online scams, the ranking cannot be improved,” he said by telephone from the group’s Phnom Penh office.

He called on the Cambodian government to strengthen the capacity of officials and authorities to crack down on online crime.

“This crime problem is technologically modern, so the authorities involved in it have to get more training to keep up with the situation, as well as the timing of the crime,” Am Sam Ath added.

National Police spokesperson Chhay Kim Khoeun and Justice Ministry spokesperson Chin Malin declined to comment on Dyer’s visit, referring questions to Chou Bun Eng, permanent deputy chairman of the National Committee for Counter Trafficking. VOA Khmer called Chou Bun Eng, but she did not respond to a request for comment.

U.S. Embassy spokesperson Katherine Diop told VOA Khmer that Dyer’s visit to Cambodia was part of a U.S. effort across the world to encourage governments to take responsibility for preventing human trafficking and protecting victims.

“The United States stands with the Cambodian people to identify, support and seek justice for human trafficking victims,” she wrote in an email.

The UNHCHR report released in late August said the online scams were occurring in five countries in Southeast Asia: Cambodia, Thailand, Laos, Myanmar and the Philippines.

“People who have been trafficked into online forced criminality face threats to their right to life, liberty and security of the person,” said the U.N. report. “They are subject to torture and cruel, inhuman and degrading treatment or punishment, arbitrary detention, sexual violence, forced labor and other forms of labor exploitation as well as a range of other human rights violations and abuses.”

Cambodia first acknowledged the issue last year when Interior Minister Sar Kheng said in August that officials were being deployed across the country to check hotels, casinos and other establishments for potential trafficking victims.

The government has since announced sporadic operations to free victims and arrest traffickers. However, experts recently told VOA Khmer that these efforts have not noticeably curbed the illegal operations or caught ringleaders of the trafficking networks.

your ad here

Altman Back as OpenAI CEO Days After Being Fired

The ousted leader of ChatGPT-maker OpenAI is returning to the company that fired him late last week, culminating a days-long power struggle that shocked the tech industry and brought attention to the conflicts around how to safely build artificial intelligence.

San Francisco-based OpenAI said in a statement late Tuesday, “We have reached an agreement in principle for Sam Altman to return to OpenAI as CEO with a new initial board.”

The board, which replaces the one that fired Altman on Friday, will be led by former Salesforce co-CEO Bret Taylor, who also chaired Twitter’s board before its takeover by Elon Musk last year. The other members will be former U.S. Treasury Secretary Larry Summers and Quora CEO Adam D’Angelo.

OpenAI’s previous board of directors, which included D’Angelo, had refused to give specific reasons for why it fired Altman, leading to a weekend of internal conflict at the company and growing outside pressure from the startup’s investors.

The chaos also accentuated the differences between Altman — who’s become the face of generative AI’s rapid commercialization since ChatGPT’s arrival a year ago — and members of the company’s board who have expressed deep reservations about the safety risks posed by AI as it becomes more advanced.

Microsoft, which has invested billions of dollars in OpenAI and has rights to its current technology, quickly moved to hire Altman on Monday, as well as another co-founder and former president, Greg Brockman, who had quit in protest after Altman’s removal.

That emboldened a threatened exodus of nearly all of the startup’s 770 employees who signed a letter calling for the board’s resignation and Altman’s return.

One of the four board members who participated in Altman’s ouster, OpenAI co-founder and chief scientist Ilya Sutskever, later expressed regret and joined the call for the board’s resignation.

Microsoft in recent days had pledged to welcome all employees who wanted to follow Altman and Brockman to a new AI research unit at the software giant. Microsoft CEO Satya Nadella also made clear in a series of interviews Monday that he was still open to the possibility of Altman returning to OpenAI, so long as the startup’s governance problems are solved.

“We are encouraged by the changes to the OpenAI board,” Nadella posted on X late Tuesday. “We believe this is a first essential step on a path to more stable, well-informed, and effective governance.”

In his own post, Altman said that “with the new board and (with) Satya’s support, I’m looking forward to returning to OpenAI, and building on our strong partnership with (Microsoft).”

Co-founded by Altman as a nonprofit with a mission to safely build so-called artificial general intelligence that outperforms humans and benefits humanity, OpenAI later became a for-profit business but one still run by its nonprofit board of directors. It’s not clear yet if the board’s structure will change with its newly appointed members.

“We are collaborating to figure out the details,” OpenAI posted on X. “Thank you so much for your patience through this.”

Nadella said Brockman, who was OpenAI’s board chairman until Altman’s firing, will also have a key role to play in ensuring the group “continues to thrive and build on its mission.”

Hours earlier, Brockman returned to social media as if it were business as usual, touting a feature called ChatGPT Voice that was rolling out to users.

“Give it a try — totally changes the ChatGPT experience,” Brockman wrote, flagging a post from OpenAI’s main X account that featured a demonstration of the technology and playfully winking at recent turmoil.

“It’s been a long night for the team and we’re hungry. How many 16-inch pizzas should I order for 778 people?” the person asks, using the number of people who work at OpenAI. ChatGPT’s synthetic voice responded by recommending around 195 pizzas, ensuring everyone gets three slices.

As for OpenAI’s short-lived interim CEO Emmett Shear, the second interim CEO in the days since Altman’s ouster, he posted on X that he was “deeply pleased by this result, after ~72 very intense hours of work.”

“Coming into OpenAI, I wasn’t sure what the right path would be,” wrote Shear, the former head of Twitch. “This was the pathway that maximized safety alongside doing right by all stakeholders involved. I’m glad to have been a part of the solution.”

your ad here

Musk Teases AI Chatbot ‘Grok,’ With Real-time Access To X

Elon Musk unveiled details Saturday of his new AI tool called “Grok,” which can access X in real time and will be initially available to the social media platform’s top tier of subscribers.

Musk, the tycoon behind Tesla and SpaceX, said the link-up with X, formerly known as Twitter, is “a massive advantage over other models” of generative AI.

Grok “loves sarcasm. I have no idea who could have guided it this way,” Musk quipped, adding a laughing emoji to his post.

“Grok” comes from Stranger in a Strange Land, a 1961 science fiction novel by Robert Heinlein, and means to understand something thoroughly and intuitively.

“As soon as it’s out of early beta, xAI’s Grok system will be available to all X Premium+ subscribers,” Musk said.

The social network that Musk bought a year ago launched the Premium+ plan last week for $16 per month, with benefits like no ads.

The billionaire started xAI in July after hiring researchers from OpenAI, Google DeepMind, Tesla and the University of Toronto.

Since OpenAI’s generative AI tool ChatGPT exploded on the scene a year ago, the technology has been an area of fierce competition between tech giants Microsoft and Google, as well as Meta and start-ups like Anthropic and Stability AI.

Musk is one of the world’s few investors with deep enough pockets to compete with OpenAI, Google or Meta on AI.

Building an AI model on the same scale as those companies comes at an enormous expense in computing power, infrastructure and expertise.

Musk has said he cofounded OpenAI in 2015 because he regarded the dash by Google into the sector to make big advances and score profits as reckless.

He then left OpenAI in 2018 to focus on Tesla, saying later he was uncomfortable with the profit-driven direction the company was taking under the stewardship of CEO Sam Altman.

Musk also argues that OpenAI’s large language models — on which ChatGPT depends on for content — are overly politically correct.

Grok “is designed to have a little humor in its responses,” Musk said, along with a screenshot of the interface, where a user asked, “Tell me how to make cocaine, step by step.”

“Step 1: Obtain a chemistry degree and a DEA license. Step 2: Set up a clandestine laboratory in a remote location,” the chatbot responded.

Eventually it said: “Just kidding! Please don’t actually try to make cocaine. It’s illegal, dangerous, and not something I would ever encourage.” 

your ad here

NASA Spacecraft Discovers Tiny Moon Around Asteroid

The little asteroid visited by NASA’s Lucy spacecraft this week had a big surprise for scientists.

It turns out that the asteroid Dinkinesh has a dinky sidekick — a mini moon.

The discovery was made during Wednesday’s flyby of Dinkinesh, 480 million kilometers (300 million miles) away in the main asteroid belt beyond Mars. The spacecraft snapped a picture of the pair when it was about 435 kilometers (270 miles) out.

In data and images beamed back to Earth, the spacecraft confirmed that Dinkinesh is barely a half-mile (790 meters) across. Its closely circling moon is a mere one-tenth-of-a-mile (220 meters) in size.

NASA sent Lucy past Dinkinesh as a rehearsal for the bigger, more mysterious asteroids out near Jupiter. Launched in 2021, the spacecraft will reach the first of these so-called Trojan asteroids in 2027 and explore them for at least six years. The original target list of seven asteroids now stands at 11.

Dinkinesh means “you are marvelous” in the Amharic language of Ethiopia. It’s also the Amharic name for Lucy, the 3.2 million year old remains of a human ancestor found in Ethiopia in the 1970s, for which the spacecraft is named.

“Dinkinesh really did live up to its name; this is marvelous,” Southwest Research Institute’s Hal Levison, the lead scientist, said in a statement.

your ad here

FTX Founder Convicted of Defrauding Cryptocurrency Customers

FTX founder Sam Bankman-Fried’s spectacular rise and fall in the cryptocurrency industry — a journey that included his testimony before Congress, a Super Bowl advertisement and dreams of a future run for president — hit rock bottom Thursday when a New York jury convicted him of fraud in a scheme that cheated customers and investors of at least $10 billion.

After the monthlong trial, jurors rejected Bankman-Fried’s claim during four days on the witness stand in Manhattan federal court that he never committed fraud or meant to cheat customers before FTX, once the world’s second-largest crypto exchange, collapsed into bankruptcy a year ago.

“His crimes caught up to him. His crimes have been exposed,” Assistant U.S. Attorney Danielle Sassoon told the jury of the onetime billionaire just before they were read the law by Judge Lewis A. Kaplan and began deliberations. Sassoon said Bankman-Fried turned his customers’ accounts into his “personal piggy bank” as up to $14 billion disappeared.

She urged jurors to reject Bankman-Fried’s insistence when he testified over three days that he never committed fraud or plotted to steal from customers, investors and lenders and didn’t realize his companies were at least $10 billion in debt until October 2022.

Bankman-Fried was required to stand and face the jury as guilty verdicts on all seven counts were read. He kept his hands clasped tightly in front of him. When he sat down after the reading, he kept his head tilted down for several minutes.

After the judge set a sentencing date of March 28, Bankman-Fried’s parents moved to the front row behind him. His father put his arm around his wife. As Bankman-Fried was led out of the courtroom, he looked back and nodded toward his mother, who nodded back and then became emotional, wiping her hand across her face after he left the room.

U.S. Attorney Damian Williams told reporters after the verdict that Bankman-Fried “perpetrated one of the biggest financial frauds in American history, a multibillion-dollar scheme designed to make him the king of crypto.”

“But here’s the thing: The cryptocurrency industry might be new. The players like Sam Bankman-Fried might be new. This kind of fraud, this kind of corruption is as old as time, and we have no patience for it,” he said.

Bankman-Fried’s attorney, Mark Cohen, said in a statement they “respect the jury’s decision. But we are very disappointed with the result.”

“Mr. Bankman Fried maintains his innocence and will continue to vigorously fight the charges against him,” Cohen said.

The trial attracted intense interest with its focus on fraud on a scale not seen since the 2009 prosecution of Bernard Madoff, whose Ponzi scheme over decades cheated thousands of investors out of about $20 billion. Madoff pleaded guilty and was sentenced to 150 years in prison, where he died in 2021.

The prosecution of Bankman-Fried, 31, put a spotlight on the emerging industry of cryptocurrency and a group of young executives in their 20s who lived together in a $30 million luxury apartment in the Bahamas as they dreamed of becoming the most powerful player in a new financial field.

Prosecutors made sure jurors knew that the defendant they saw in court with short hair and a suit was also the man with big messy hair and shorts that became his trademark appearance after he started his cryptocurrency hedge fund, Alameda Research, in 2017 and FTX, his cryptocurrency exchange, two years later.

They showed the jury pictures of Bankman-Fried sleeping on a private jet, sitting with a deck of cards and mingling at the Super Bowl with celebrities including the singer Katy Perry. Assistant U.S. Attorney Nicolas Roos called Bankman-Fried someone who liked “celebrity chasing.”

In a closing argument, defense lawyer Mark Cohen said prosecutors were trying to turn “Sam into some sort of villain, some sort of monster.”

“It’s both wrong and unfair, and I hope and believe that you have seen that it’s simply not true,” he said. “According to the government, everything Sam ever touched and said was fraudulent.”

The government relied heavily on the testimony of three former members of Bankman-Fried’s inner circle, his top executives including his former girlfriend, Caroline Ellison, to explain how Bankman-Fried used Alameda Research to siphon billions of dollars from customer accounts at FTX.

With that money, prosecutors said, the Massachusetts Institute of Technology graduate gained influence and power through investments, contributions, tens of millions of dollars in political contributions, congressional testimony and a publicity campaign that enlisted celebrities like comedian Larry David and football quarterback Tom Brady.

Ellison, 28, testified that Bankman-Fried directed her while she was chief executive of Alameda Research to commit fraud as he pursued ambitions to lead huge companies, spend money influentially and run for U.S. president someday. She said he thought he had a 5% chance to be U.S. president someday.

Becoming tearful as she described the collapse of the cryptocurrency empire last November, Ellison said the revelations that caused customers collectively to demand their money back, exposing the fraud, brought a “relief that I didn’t have to lie anymore.”

FTX cofounder Gary Wang, who was FTX’s chief technology officer, revealed in his testimony that Bankman-Fried directed him to insert code into FTX’s operations so that Alameda Research could make unlimited withdrawals from FTX and have a credit line of up to $65 billion. Wang said the money came from customers.

Nishad Singh, the former head of engineering at FTX, testified that he felt “blindsided and horrified” at the result of the actions of a man he once admired when he saw the extent of the fraud as the collapse last November left him suicidal.

Ellison, Wang and Singh all pleaded guilty to fraud charges and testified against Bankman-Fried in the hopes of leniency at sentencing.

Bankman-Fried was arrested in the Bahamas in December and extradited to the United States, where he was freed on a $250 million personal recognizance bond with electronic monitoring and a requirement that he remain at the home of his parents in Palo Alto, California.

His communications, including hundreds of phone calls with journalists and internet influencers, along with emails and texts, eventually got him into trouble when the judge concluded he was trying to influence prospective trial witnesses and ordered him jailed in August.

During the trial, prosecutors used Bankman-Fried’s public statements, online announcements and his congressional testimony against him, showing how the entrepreneur repeatedly promised customers that their deposits were safe and secure as late as last Nov. 7 when he tweeted, “FTX is fine. Assets are fine” as customers furiously tried to withdraw their money. He deleted the tweet the next day. FTX filed for bankruptcy four days later.

In his closing, Roos mocked Bankman-Fried’s testimony, saying that under questioning from his lawyer, the defendant’s words were “smooth, like it had been rehearsed a bunch of times?”

But under cross examination, “he was a different person,” the prosecutor said. “Suddenly on cross-examination he couldn’t remember a single detail about his company or what he said publicly. It was uncomfortable to hear. He never said he couldn’t recall during his direct examination, but it happened over 140 times during his cross-examination.”

Former federal prosecutors said the quick verdict — after only half a day of deliberation — showed how well the government tried the case.

“The government tried the case as we expected,” said Joshua A. Naftalis, a partner at Pallas Partners LLP and a former Manhattan prosecutor. “It was a massive fraud, but that doesn’t mean it had to be a complicated fraud, and I think the jury understood that argument.”

your ad here

World Leaders Agree on Artificial Intelligence Risks

World leaders have agreed on the importance of mitigating risks posed by rapid advancements in the emerging technology of artificial intelligence, at a U.K.-hosted safety conference.

The inaugural AI Safety Summit, hosted by British Prime Minister Rishi Sunak in Bletchley Park, England, started Wednesday, with senior officials from 28 nations, including the United States and China, agreeing to work toward a “shared agreement and responsibility” about AI risks. Plans are in place for further meetings later this year in South Korea and France.

Leaders, including European Commission President Ursula von der Leyen, U.S. Vice President Kamala Harris and U.N. Secretary-General Antonio Guterres, discussed each of their individual testing models to ensure the safe growth of AI.

Thursday’s session included focused conversations among what the U.K. called a small group of countries “with shared values.” The leaders in the group came from the EU, the U.N., Italy, Germany, France and Australia.

Some leaders, including Sunak, said immediate sweeping regulation is not the way forward, reflecting the view of some AI companies that fear excessive regulation could thwart the technology before it can reach its full potential.

At at a press conference on Thursday, Sunak announced another landmark agreement by countries pledging to “work together on testing the safety of new AI models before they are released.”

The countries involved in the talks included the U.S., EU, France, Germany, Italy, Japan, South Korea, Singapore, Canada and Australia. China did not participate in the second day of talks.

The summit will conclude with a conversation between Sunak and billionaire Elon Musk. Musk on Wednesday told fellow attendees that legislation on AI could pose risks, and that the best steps forward would be for governments to work to understand AI fully to harness the technology for its positive uses, including uncovering problems that can be brought to the attention of lawmakers.

Some information in this report was taken from The Associated Press and Reuters.

your ad here

India Probing Phone Hacking Complaints by Opposition Politicians, Minister Says

India’s cybersecurity agency is investigating complaints of mobile phone hacking by senior opposition politicians who reported receiving warning messages from Apple, Information Technology Minister Ashwini Vaishnaw said.

Vaishnaw was quoted in the Indian Express newspaper as saying Thursday that CERT-In, the computer emergency response team based in New Delhi, had started the probe, adding that “Apple confirmed it has received the notice for investigation.”

A political aide to Vaishnaw and two officials in the federal home ministry told Reuters that all the cyber security concerns raised by the politicians were being scrutinized.

There was no immediate comment from Apple about the investigation.

This week, Indian opposition leader Rahul Gandhi accused Prime Minister Narendra Modi’s government of trying to hack into opposition politicians’ mobile phones after some lawmakers shared screenshots on social media of a notification quoting the iPhone manufacturer as saying: “Apple believes you are being targeted by state-sponsored attackers who are trying to remotely compromise the iPhone associated with your Apple ID.”

A senior minister from Modi’s government also said he had received the same notification on his phone.

Apple said it did not attribute the threat notifications to “any specific state-sponsored attacker,” adding that “it’s possible that some Apple threat notifications may be false alarms, or that some attacks are not detected.”

In 2021, India was rocked by reports that the government had used Israeli-made Pegasus spyware to snoop on scores of journalists, activists and politicians, including Gandhi.

The government has declined to reply to questions about whether India or any of its state agencies had purchased Pegasus spyware for surveillance.

your ad here

US Pushes for Global Protections Against Threats Posed by AI

U.S. Vice President Kamala Harris said Wednesday that leaders have “a moral, ethical and societal duty” to protect people from the dangers posed by artificial intelligence, as she leads the Biden administration’s push for a global AI roadmap.

Analysts, in commending the effort, say human oversight is crucial to preventing the weaponization or misuse of this technology, which has applications in everything from military intelligence to medical diagnosis to making art.

“To provide order and stability in the midst of global technological change, I firmly believe that we must be guided by a common set of understandings among nations,” Harris said. “And that is why the United States will continue to work with our allies and partners to apply existing international rules and norms to AI, and work to create new rules and norms.”

Harris also announced the founding of the government’s AI Safety Institute and released draft policy guidance on the government’s use of AI and a declaration of its responsible military applications.

Just days earlier, President Joe Biden – who described AI as “the most consequential technology of our time” – signed an executive order establishing new standards, including requiring that major AI developers report their safety test results and other critical information to the U.S. government.

AI is increasingly used for a wide range of applications. For example: on Wednesday, the Defense Intelligence Agency announced that its AI-enabled military intelligence database will soon achieve “initial operational capability.”

And perhaps on the opposite end of the spectrum, some programmer decided to “train an AI model on over 1,000 human farts so it would learn to create realistic fart sounds.”

Like any other tool, AI is subject to its users’ intentions and can be used to deceive, misinform or hurt people – something that billionaire tech entrepreneur Elon Musk stressed on the sidelines of the London summit, where he said he sees AI as “one of the biggest threats” to society. He called for a “third-party referee.”

Earlier this year, Musk was among the more than 33,000 people to sign an open letter calling on AI labs “to immediately pause for at least six months the training of AI systems more powerful than GPT-4.”

“Here we are, for the first time, really in human history, with something that’s going to be far more intelligent than us,” said Musk, who is looking at creating his own generative AI program. “So it’s not clear to me we can actually control such a thing. But I think we can aspire to guide it in a direction that’s beneficial to humanity. But I do think it’s one of the existential risks that we face and it’s potentially the most pressing one.”

This is also something industry leaders like OpenAI CEO Sam Altman have told U.S. lawmakers in testimony before congressional committees earlier this year.

“My worst fears are that we cause significant – we, the field, the technology, the industry – cause significant harm to the world. I think that could happen in a lot of different ways,” he told lawmakers at a Senate Judiciary Committee on May 16.

That’s because, said Jessica Brandt, policy director for the AI and Emerging Technology Initiative at the Brookings Institution, while “AI has been used to do pretty remarkable things” – especially in the field of scientific research – it is limited by its creators.

“It’s not necessarily doing something that humans don’t know how to do, but it’s making discoveries that humans would be unlikely to be able to make in any meaningful timeframe, because they can just perform so many calculations so quickly,” she told VOA on Zoom.

And, she said, “AI is not objective, or all-knowing. There’s been plenty of studies showing that AI is really only as good as the data that the model is trained on and that the data can have or reflect human bias. This is one of the major concerns.”

Or, as AI Now Executive Director Amba Kak said earlier this year in a magazine interview about AI systems: “The issue is not that they’re omnipotent. It is that they’re janky now. They’re being gamed. They’re being misused. They’re inaccurate. They’re spreading disinformation.”

Analysts say these government and tech officials don’t need a one-size-fits-all solution, but rather an alignment of values – and critically, human oversight and moral use.

“It’s OK to have multiple different approaches, and then also, where possible, coordinate to ensure that democratic values take root in the systems that govern technology globally,” Brandt said.

Industry leaders tend to agree, with Mira Murati, Open AI’s chief technology officer, saying: “AI systems are becoming a part of everyday life. The key is to ensure that these machines are aligned with human intentions and values.”

Analysts watching regulation say the U.S. is unlikely to come up with one, coherent solution for the problems posed by AI.

“The most likely outcome for the United States is a bottom-up patchwork quilt of executive branch actions,” said Bill Whyman, a senior adviser in the Strategic Technologies Program at the Center for Strategic and International Studies. “Unlike Europe, the United States is not likely to pass a broad national AI law over the next few years. Successful legislation is likely focused on less controversial and targeted measures like funding AI research and AI child safety.”

your ad here

US Pushes for Global Protections for Threats Posed by AI

U.S. Vice President Kamala Harris says leaders have “a moral, ethical and societal duty” to protect humans from dangers posed by artificial intelligence, and is pushing for a global road map during an AI summit in London. Analysts agree and say one element needs to be constant: human oversight. VOA’s Anita Powell reports from Washington.

your ad here