×
Expert: The Trump Doctrine Aims to Undermine ‘China’s Project to Destroy America’

Expert: The Trump Doctrine Aims to Undermine ‘China’s Project to Destroy America’

WASHINGTON – The war with Iran is having a worldwide impact that’s being felt all the way to China. That’s because China imports most of its oil from Tehran. But there is debate over what impact the war is having beyond oil. One expert sees the Trump Doctrine undermining China’s plan for world domination. 

As the U.S. military bombards Iranian targets, the world watches and waits to see how it will all end. China, the top adversary of the United States, has a lot riding on the outcome. 

China expert Gordon Chang, author of Plan Red: China’s Project to Destroy America, says while oil gets most of the attention, he believes the broader implications are much larger. Chang sees a more sinister force between the two countries because China helped Iran build key parts of its nuclear program by working through Pakistan’s “father of the nuclear bomb.”

***Please sign up for CBN Newsletters and download the CBN News app to receive the latest news from a Christian perspective.***
 
Chang said, “For instance, Iran obtained its centrifuges through the A.Q. Khan black-market network in Pakistan. But A.Q. Khan was really a pawn for China, because he was selling Chinese technology—and Beijing knew what he was doing. So you can say that Iran’s nuclear program is essentially China’s program transplanted to the Gulf region.”

“Iran also has many Chinese weapons. Inside Iranian systems, you find Chinese components, especially microchips, either made in China or shipped to Iran through Chinese intermediaries,” he further explained.
 
The Trump administration says Iran’s refusal to stop its nuclear ambitions by enriching uranium and building more ballistic missiles led to the current conflict. Chang says that defiance is due to Iran’s alliance with China to keep conflict alive against the U.S. and Israel, using proxies throughout the Middle East.
 
“China is trying to convince the world it is a responsible partner on nonproliferation. In reality, China has been the world’s leading proliferator of nuclear weapons technology, much of it to Iran. And it wasn’t just centrifuges. It was also Chinese warhead designs. This is a comprehensive program by the Chinese regime to use the spread of that technology to advance its own goals.”
 
According to Chang, China seeks to replace the United States and dominate both economically and militarily as the world’s top superpower. 

MORE: China Is Gaming the US ‘Birthright Citizenship’ System to Plan an ‘Invisible Coup’

“The President will soon visit Xi Jinping in China. What will these two leaders need to discuss regarding China’s appetite for disrupting its neighbors—including Taiwan—and its growing threat to the United States?”
 
“I think President Trump needs to say, in no uncertain terms, to the Chinese leader that the United States will defend Taiwan, and we will defend ourselves, our friends, and our allies,” Chang said.

“Xi Jinping is very arrogant. Until the end of last year, he believed he could push President Trump around. That changed on January 3, when Trump removed Maduro and his wife from Caracas, and then with the attack on Iran. But we are still dealing with a very arrogant Chinese leadership right now.”
 
President Trump had planned to meet with President Xi in March. Due to the military conflict with Iran, he rescheduled his visit to May.

Source link
#Expert #Trump #Doctrine #Aims #Undermine #Chinas #Project #Destroy #America

岩手 大槌町 山林火災は住宅近く迫る 1000人規模で消火 | NHKニュース岩手県大槌町の山林火災は25日で発生から4日目となりますが、延焼が続いています。町によりますと複数の地区にある住宅の近くまで火が迫っているということで、1000人以上の規模での懸命な消火活動が行われています。#岩手 #大槌町 #山林火災は住宅近く迫る #1000人規模で消火 #NHKニュースNHK,ニュース,NHK ONE,火災,岩手県,一覧

OpenAI CEO Sam Altman has apologized to the Canadian town of Tumbler Ridge following a February mass shooting that left eight dead. 

Altman said he was “deeply sorry” the company didn’t alert the police about the shooter’s troubling ChatGPT accounts.

Britich Colombia Primier David Eby called the apology “necessary, and yet grossly insufficient.”

How did OpenAI fail to act?

An 18-year-old transgender woman killed her mother and stepbrother at home on February 10, before going to a local secondary school and opening fire. She killed five children and a teacher, then took her own life.

After the attack, OpenAI said it had identified the suspect’s account through its abuse detection systems and banned it in June, eight months before the shooting.

The ChatGPT developer said it did not report the account to Canadian police at the time, as the activity did not meet its threshold for referral to law enforcement.

“I am deeply sorry that we did not alert law enforcement to the account that was banned in June,” Altman said. “While I know words can never be enough, I believe an apology is necessary to recognize the harm and irreversible loss your community has suffered.” 

How does ChatGPT report suspected violance?

OpenAI says it uses automated moderation systems that scan content in real time. Accounts can be restricted or banned for violating the rules. Violations include sexual exploitation, support of self-harm and suicide, and promotion of violence and harm.

In serious cases, systems are designed to flag high-risk behavior for human review. If a credible threat is identified, the company may share relevant account data with law enforcement.

Following the attack, Canadian officials summoned OpenAI’s safety team and warned of regulation actions if changes were not made. The company said it would tighten its safety measures and had created a direct contact channel with police.

In the letter, Altman said the company is committed to find ways to prevent similar tragedies. “Going forward, our focus will continue to be on working with all levels of government to help ensure something like this never happens again,” he said. 

The family of a girl who was seriously injured in the shooting has filed a negligence lawsuit against the US tech giant.

Is your AI private? OpenAI and the Canadian school shooting

Edited by: Wesley Dockery 

#OpenAI #apologizes #reporting #Canada #mass #shooter">OpenAI apologizes for not reporting Canada mass shooterOpenAI CEO Sam Altman has apologized to the Canadian town of Tumbler Ridge following a February mass shooting that left eight dead. 

Altman said he was “deeply sorry” the company didn’t alert the police about the shooter’s troubling ChatGPT accounts.

Britich Colombia Primier David Eby called the apology “necessary, and yet grossly insufficient.”

How did OpenAI fail to act?

An 18-year-old transgender woman killed her mother and stepbrother at home on February 10, before going to a local secondary school and opening fire. She killed five children and a teacher, then took her own life.

After the attack, OpenAI said it had identified the suspect’s account through its abuse detection systems and banned it in June, eight months before the shooting.

The ChatGPT developer said it did not report the account to Canadian police at the time, as the activity did not meet its threshold for referral to law enforcement.

“I am deeply sorry that we did not alert law enforcement to the account that was banned in June,” Altman said. “While I know words can never be enough, I believe an apology is necessary to recognize the harm and irreversible loss your community has suffered.” 

How does ChatGPT report suspected violance?

OpenAI says it uses automated moderation systems that scan content in real time. Accounts can be restricted or banned for violating the rules. Violations include sexual exploitation, support of self-harm and suicide, and promotion of violence and harm.

In serious cases, systems are designed to flag high-risk behavior for human review. If a credible threat is identified, the company may share relevant account data with law enforcement.

Following the attack, Canadian officials summoned OpenAI’s safety team and warned of regulation actions if changes were not made. The company said it would tighten its safety measures and had created a direct contact channel with police.

In the letter, Altman said the company is committed to find ways to prevent similar tragedies. “Going forward, our focus will continue to be on working with all levels of government to help ensure something like this never happens again,” he said. 

The family of a girl who was seriously injured in the shooting has filed a negligence lawsuit against the US tech giant.

Is your AI private? OpenAI and the Canadian school shootingTo view this video please enable JavaScript, and consider upgrading to a web browser that supports HTML5 video

Edited by: Wesley Dockery 
#OpenAI #apologizes #reporting #Canada #mass #shooter

February mass shooting that left eight dead. 

Altman said he was “deeply sorry” the company didn’t alert the police about the shooter’s troubling ChatGPT accounts.

Britich Colombia Primier David Eby called the apology “necessary, and yet grossly insufficient.”

How did OpenAI fail to act?

An 18-year-old transgender woman killed her mother and stepbrother at home on February 10, before going to a local secondary school and opening fire. She killed five children and a teacher, then took her own life.

After the attack, OpenAI said it had identified the suspect’s account through its abuse detection systems and banned it in June, eight months before the shooting.

The ChatGPT developer said it did not report the account to Canadian police at the time, as the activity did not meet its threshold for referral to law enforcement.

“I am deeply sorry that we did not alert law enforcement to the account that was banned in June,” Altman said. “While I know words can never be enough, I believe an apology is necessary to recognize the harm and irreversible loss your community has suffered.” 

How does ChatGPT report suspected violance?

OpenAI says it uses automated moderation systems that scan content in real time. Accounts can be restricted or banned for violating the rules. Violations include sexual exploitation, support of self-harm and suicide, and promotion of violence and harm.

In serious cases, systems are designed to flag high-risk behavior for human review. If a credible threat is identified, the company may share relevant account data with law enforcement.

Following the attack, Canadian officials summoned OpenAI’s safety team and warned of regulation actions if changes were not made. The company said it would tighten its safety measures and had created a direct contact channel with police.

In the letter, Altman said the company is committed to find ways to prevent similar tragedies. “Going forward, our focus will continue to be on working with all levels of government to help ensure something like this never happens again,” he said. 

The family of a girl who was seriously injured in the shooting has filed a negligence lawsuit against the US tech giant.

Is your AI private? OpenAI and the Canadian school shooting

Edited by: Wesley Dockery 

#OpenAI #apologizes #reporting #Canada #mass #shooter">OpenAI apologizes for not reporting Canada mass shooter

OpenAI CEO Sam Altman has apologized to the Canadian town of Tumbler Ridge following a February mass shooting that left eight dead. 

Altman said he was “deeply sorry” the company didn’t alert the police about the shooter’s troubling ChatGPT accounts.

Britich Colombia Primier David Eby called the apology “necessary, and yet grossly insufficient.”

How did OpenAI fail to act?

An 18-year-old transgender woman killed her mother and stepbrother at home on February 10, before going to a local secondary school and opening fire. She killed five children and a teacher, then took her own life.

After the attack, OpenAI said it had identified the suspect’s account through its abuse detection systems and banned it in June, eight months before the shooting.

The ChatGPT developer said it did not report the account to Canadian police at the time, as the activity did not meet its threshold for referral to law enforcement.

“I am deeply sorry that we did not alert law enforcement to the account that was banned in June,” Altman said. “While I know words can never be enough, I believe an apology is necessary to recognize the harm and irreversible loss your community has suffered.” 

How does ChatGPT report suspected violance?

OpenAI says it uses automated moderation systems that scan content in real time. Accounts can be restricted or banned for violating the rules. Violations include sexual exploitation, support of self-harm and suicide, and promotion of violence and harm.

In serious cases, systems are designed to flag high-risk behavior for human review. If a credible threat is identified, the company may share relevant account data with law enforcement.

Following the attack, Canadian officials summoned OpenAI’s safety team and warned of regulation actions if changes were not made. The company said it would tighten its safety measures and had created a direct contact channel with police.

In the letter, Altman said the company is committed to find ways to prevent similar tragedies. “Going forward, our focus will continue to be on working with all levels of government to help ensure something like this never happens again,” he said. 

The family of a girl who was seriously injured in the shooting has filed a negligence lawsuit against the US tech giant.

Is your AI private? OpenAI and the Canadian school shooting

Edited by: Wesley Dockery 

#OpenAI #apologizes #reporting #Canada #mass #shooter

Post Comment