Connect with us


North Korea’s about-face complicates South Korea’s pitch to Trump | World

North Korean leader Kim Jong Un and South Korean President Moon Jae-in deliver a statement at the truce village of Panmunjom inside the demilitarised zone separating the two Koreas, South Korea, April 27, 2018. Photo: Reuters 

SEOUL: South Korean President Moon Jae-in has been riding high after a string of successes with North Korea, seeing his popularity surge after his historic meeting with Kim Jong Un as he looks ahead to the US-North Korea summit he helped engineer next month.

But just days before a key meeting with US President Donald Trump, Pyongyang has made Moon’s role as a diplomatic go-between much harder, calling his government “ignorant and incompetent” and pulling out of inter-Korean talks.

Moon heads to Washington next week with the daunting task of crafting a unified approach between the two allies ahead of the June 12 meeting between Trump and Kim in Singapore.

While publicly supporting Trump’s campaign of sanctions and military threats, Moon has long called for reconciliation with the North and has pushed Washington to engage with Pyongyang’s overtures.

North Korea’s threat this week to scrap the Trump summit if Washington continues to press for unilateral denuclearisation and complaints about US-South Korea military drills has done little to advance Moon’s hope for a peace deal.

“Moon did an impressive job building up political capital with both leaders — with Trump throughout 2017 and Kim more in 2018 — and has spent some of it to create a situation that brings North Korea and the US to the table,” said Andray Abrahamian, adjunct fellow at the Griffith Asia Institute. “He’s still in a position to act as a go-between, but only as far as the other two parties take bridgeable positions.”

Complicating factors 

The North’s outburst aimed at Moon’s government appeared partially driven by anger over South Korea allowing Thae Yong Ho, a former North Korean diplomat to Britain who defected to the South in 2016, to hold a press conference on Monday at the South Korean National Assembly for the publication of his memoir.

In his memoir, “Password from the Third Floor”, Thae describes North Korean leader Kim as “impatient, impulsive and violent”.

In rambling remarks in the White House’s Oval Office in which he also sharply criticised China over trade, Trump said that as far as he knew the meeting with Kim was still on track, but that the North Korean leader was possibly being influenced by Beijing after two recent visits he made there.

North Korea’s renewed ties with China “certainly has emboldened” Kim by making him less reliant on a successful summit with Trump, but Beijing is unlikely to be directing the latest outbursts, said Zhao Tong, a North Korea expert at the Carnegie-Tsinghua Center in Beijing.

“I am skeptical about the speculation that China was behind Kim Jong Un’s renewed hard line,” Zhao said. “I believe China wants North Korea and the US to reach a deal, which would make it easier for China to enhance economic cooperation with North Korea and better facilitate North Korea’s strategic transition to economic development and open-up.”

‘Bridge for success’ 

Moon’s administration quickly accepted North Korean leader Kim Jong Un’s New Year’s proposal to ease tensions, arranged a historic inter-Korean summit, and passed on Kim’s invitation to meet with Trump.

North Korea, meanwhile, has pledged to publicly dismantle its nuclear test site next week.

South Korean officials said they still believe the nuclear site will be dismantled, and said they had not seen any changes in their lines of communication with the North Koreans.

Moon is scheduled to meet with Trump on Tuesday in Washington.

Despite professed unity, Trump has often taken a harder line on North Korea than Moon, and the US president has repeatedly criticised South Korea over trade while questioning the usefulness of the alliance.

South Korea’s presidential Blue House, however, says it hopes Moon’s meeting with Trump will help set the stage for more progress with North Korea.

“As this US-South Korea summit is taking place three weeks ahead of the US-North Korea summit, we hope it will act as a bridge for the success of the US-North Korea summit,” said Nam Gwan-pyo, second deputy director of the Blue House’s National Security Office.

Source link

Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

child abuse

Judge receives death threats in child abuse case

LOS ANGELES: A US judge has received death threats after granting bail to five suspects in a case involving nearly a dozen allegedly malnourished children found at a remote compound in New Mexico.

Judge Sarah Backus granted the three men and two women — described by authorities as extremists — bail on Monday on grounds that prosecutors had failed to present clear and convincing evidence that they posed a threat.

Her decision, linked to bail reforms adopted in the western US state two years ago, stunned many in the community and prompted death threats against her and a barrage of abusive messages on social media.

“One caller said she wished someone would come and smash the judge’s head,” Barry Massey, a court spokesman said.

Another caller said he “wished her throat were slit.”

The Taos County courthouse was closed to visitors early Tuesday as a result of the threats.

Prosecutors say the suspects in the case had trained some of the children — aged between one and 15 — to use weapons and carry out school shootings.

Police raided the compound on August 3 while investigating the disappearance of the disabled child of one of the suspects, Siraj Ibn Wahhaj.

The body of a boy, believed to be the three-year-old child missing for more than a year, was later found at the compound, buried under human waste.

Authorities said the raid was carried out after the discovery of a web posting from one of the women inside the compound that said the group was starving.

Local officials said the adults and children found at the makeshift compound looked malnourished and lived in squalor.

The five suspects released on bail have been charged with child abuse.



Source link

Continue Reading

bridge collapse

Anger grows in Italy as bridge toll hits 39

GENOA: Italy’s government on Wednesday blamed the firm that operated the collapsed Genoa bridge for the disaster in which at least 39 people died, as it announced a state of emergency in the city.

Search operations, meanwhile, were due to enter a second night with rescuers digging through mountains of crushed concrete.

A vast span of the Morandi bridge caved in during a heavy rainstorm in the northern port city on Tuesday, sending about 35 cars and several trucks plunging 45 metres (150 feet) onto railway tracks below.

Children aged eight, 12 and 13 were among the dead, Interior Minister Matteo Salvini said, adding that more people were still missing. Sixteen people were injured.

The driver of a green lorry left precariously close to the edge told Italian media how he had escaped the “hell” of the bridge collapse.

“It was raining very hard and it wasn’t possible to go very fast,” he told the Corriere della Sera daily.

“When a car overtook me I slowed down… (then) at a certain moment everything shook. The car in front of me disappeared and seemed to be swallowed up by the clouds. I Looked up and saw the bridge pylon fall,” he said.

“Instinctively, finding myself in front of the void, I put the van into reverse, to escape this hell,” he added.

Three Chileans, who live in Italy, and four French nationals were also killed.

The tragedy has focussed anger on the structural problems that have dogged the decades old Morandi bridge and the private sector firm Autostrade per l’Italia, which is currently in charge of operating and maintaining swathes of the country’s motorways.

Deputy prime minister Luigi Di Maio said the tragedy “could have been avoided”.

“Autostrade should have done maintenance and didn’t do it,” he alleged.

Prime Minister Giuseppe Conte also confirmed that his government would push to revoke the company’s contract for the A10 motorway, which includes the bridge, while Transport Minister Danilo Toninelli said the company should be fined up to 150 million euros ($170 million).

The firm, which said the bridge had been undergoing maintenance work, however, released a statement refuting accusations of underfunding of motorway infrastructure.

“In the last five years (2012-2017) the company’s investment in the security, maintenance and strengthening of the network has been over one billion euros a year,” it said.

‘I went down with the car’

Survivors recounted the heart-stopping moment when the bridge buckled, tossing vehicles and hunks of concrete into the abyss.

Davide Capello, a former goalkeeper for Italian Serie A club Cagliari, plunged with his car but was unscathed.

“I was driving along the bridge, and at a certain point I saw the road in front of me collapse, and I went down with the car,” he told TV news channel Sky TG24.

As cars and trucks tumbled off the bridge, Afifi Idriss, 39, a Moroccan truck driver, just managed to stop in time.

“I saw the green lorry in front of me stop and then reverse so I stopped too, locked the truck and ran,” he told AFP.

While around a dozen apartment blocks that stand in the shadow of the viaduct were largely spared the impact of the falling concrete, the Liguria regional government said some 634 people had been evacuated.

Interior Minister Matteo Salvini said the homes would have to be pulled down.

 ‘A tragedy waiting to happen’

The incident is the latest in a string of bridge collapses in Italy, a country prone to damage from seismic activity but where infrastructure generally is showing the effects of a faltering economy.

The Morandi viaduct, completed in 1967, spans dozens of railway lines.

The bridge has been riddled with structural problems since its construction, which has led to expensive maintenance and severe criticism from engineering experts.

On Tuesday engineering website “” called it “a tragedy waiting to happen”.

Conte also announced after a cabinet meeting Wednesday that a national day of mourning was being planned.

There would also be a 12-month state of emergency in Genoa, he said, with five million euros of funds going into recovery work.



Source link

Continue Reading


Why Facebook is losing the war on hate speech in Myanmar

In April, Facebook founder Mark Zuckerberg told US senators that the social media site was hiring dozens more Burmese speakers to review hate speech posted in Myanmar. The situation was dire.

Some 700,000 members of the Rohingya community had recently fled the country amid a military crackdown and ethnic violence. In March, a United Nations investigator said Facebook was used to incite violence and hatred against the Muslim minority group. The platform, she said, had “turned into a beast.”

Four months after Zuckerberg’s pledge to act, here is a sampling of posts from Myanmar that were viewable this month on Facebook:

One user posted a restaurant advertisement featuring Rohingya-style food. “We must fight them the way Hitler did the Jews, damn kalars!” the person wrote, using a pejorative for the Rohingya. That post went up in December 2013.

Another post showed a news article from an army-controlled publication about attacks on police stations by Rohingya militants. “These non-human kalar dogs, the Bengalis, are killing and destroying our land, our water and our ethnic people,” the user wrote. “We need to destroy their race.” That post went up last September, as the violence against the Rohingya peaked.


A third user shared a blog item that pictures a boatload of Rohingya refugees landing in Indonesia. “Pour fuel and set fire so that they can meet Allah faster,” a commenter wrote. The post appeared 11 days after Zuckerberg’s Senate testimony.

The remarks are among more than 1,000 examples Reuters found of posts, comments, images and videos attacking the Rohingya or other Myanmar Muslims that were on Facebook as of last week. Almost all are in the main local language, Burmese. The anti-Rohingya and anti-Muslim invective analysed for this article – which was collected by Reuters and the Human Rights Center at UC Berkeley School of Law – includes material that’s been up on Facebook for as long as six years.

The poisonous posts call the Rohingya or other Muslims dogs, maggots and rapists, suggest they be fed to pigs, and urge they be shot or exterminated. The material also includes crudely pornographic anti-Muslim images. The company’s rules specifically prohibit attacking ethnic groups with “violent or dehumanising speech” or comparing them to animals. Facebook also has long had a strict policy against pornographic content.

The use of Facebook to spread hate speech against the Rohingya in the Buddhist-majority country has been widely reported by the U.N. and others. Now, a Reuters investigation gives an inside look at why the company has failed to stop the problem.

For years, Facebook – which reported net income of $15.9 billion (£12.5 billion) in 2017 – devoted scant resources to combat hate speech in Myanmar, a market it dominates and in which there have been regular outbreaks of ethnic violence. In early 2015, there were only two people at Facebook who could speak Burmese reviewing problematic posts. Before that, most of the people reviewing Burmese content spoke English.

To this day, the company continues to rely heavily on users reporting hate speech in part because its systems struggle to interpret Burmese text.

Even now, Facebook doesn’t have a single employee in the country of some 50 million people. Instead, it monitors hate speech from abroad. This is mainly done through a secretive operation in Kuala Lumpur that’s outsourced to Accenture, the professional services firm, and codenamed “Project Honey Badger.”

According to people familiar with the matter, the project, which handles many Asian countries, hired its first two Burmese speakers, who were based in Manila, just three years ago. As of June, Honey Badger had about 60 people reviewing reports of hate speech and other content posted by Myanmar’s 18 million active Facebook users. Facebook itself in April had three full-time Burmese speakers at a separate monitoring operation at its international headquarters in Dublin, according to a former employee.

Honey Badger employees typically sign one-year renewable contracts and agree not to divulge that the client is Facebook. Reuters interviewed more than a half-dozen former monitors who reviewed Southeast Asian content.

A Facebook official said outsourcing its content monitoring is more efficient because the companies it uses are specialists in ramping up such operations. He declined to disclose how many Burmese speakers the company has on the job worldwide, saying it was “impossible to know, to be definitive on that.”

“It’s not enough,” he added.

For many people in this emerging economy, Facebook is the internet: It’s so dominant, it’s the only site they use online. Yet, the company ignored repeated warnings as far back as 2013 that it faced trouble.

Researchers and human rights activists say they cautioned Facebook for years that its platform was being used in Myanmar to promote racism and hatred of Muslims, in particular the Rohingya.

“They were warned so many times,” said David Madden, a tech entrepreneur who worked in Myanmar. He said he told Facebook officials in 2015 that its platform was being exploited to foment hatred in a talk he gave at its headquarters in Menlo Park, California. About a dozen Facebook people attended the meeting in person, including Mia Garlick, now the company’s director of Asia Pacific policy, he said. Others joined via video. “It couldn’t have been presented to them more clearly, and they didn’t take the necessary steps,” Madden said.

In a statement, Garlick told Reuters: “We were too slow to respond to the concerns raised by civil society, academics and other groups in Myanmar. We don’t want Facebook to be used to spread hatred and incite violence. This is true around the world, but it is especially true in Myanmar where our services can be used to amplify hate or exacerbate harm against the Rohingya.”

She added that Facebook is focussed on addressing challenges that are unique to Myanmar “through a combination of people, technology, policies and programs.” The company also said it has banned several “hate figures and organizations” on Facebook in Myanmar.

Facebook’s struggles in Myanmar are among much broader problems it faces. Zuckerberg’s congressional testimony in April primarily focussed on the company’s mishandling of user data, whether it censors conservative views and Russia’s exploitation of Facebook to meddle in the 2016 U.S. presidential election.

Of all of Facebook’s travails, though, Myanmar may be the bloodiest. The Myanmar military stands accused by the U.N. of having conducted a brutal campaign of killings, mass rape, arson and ethnic cleansing against the Rohingya. The government denies the allegations.

The social media giant doesn’t make public its data on hate speech in Myanmar. It says it has 2.2 billion global users and each week receives millions of user reports from around the world about problematic content.

In compiling examples of hate speech for this article, Reuters found some that Facebook subsequently removed. But the vast majority remained online as of early August.

After Reuters alerted Facebook to some of the derogatory posts included in this story, the company said it removed them. “All of it violated our policies,” it said.

Reuters itself sometimes flags to Facebook threats posted on the platform against its reporters. These include the Burmese journalists Wa Lone and Kyaw Soe Oo, who are on trial in Myanmar on charges of violating a state secrets law. The two were arrested in December while reporting on the massacre of 10 Rohingya men and have received a deluge of death threats on social media over their story. Facebook has removed such content several times at the news agency’s request.



Myanmar emerged from decades of military rule in 2011, but religious violence has marred its transition to democracy. In 2012, clashes in Rakhine State between ethnic Rakhine, who are Buddhists, and the Rohingya killed scores of people and left 140,000 displaced – mostly Muslims.

Facebook’s extraordinary dominance in Myanmar began taking root around the same time. But not by design.

As recently as six years ago, Myanmar was one of the least connected countries on earth. In 2012, only 1.1 percent of the population used the internet and few people had telephones, according to the International Telecommunication Union, a U.N. agency. The junta that had ruled the country for decades kept citizens isolated.

That all changed in 2013, when a quasi-civilian government oversaw the deregulation of telecommunications. The state-owned phone company suddenly faced competition from two foreign mobile-phone entrants from Norway and Qatar.

The price of SIM cards dropped from more than $200 to as little as $2 and people purchased them in droves. By 2016, nearly half the population had mobile phone subscriptions, according to GSMA Intelligence, the research arm of the industry’s trade association. Most purchased smartphones with internet access.

One app went viral: Facebook. Many saw it as an all-in-one solution – offering a messaging system, news, and videos and other entertainment. It also became a status symbol, said Chris Tun, a former Deloitte consultant who advised the government. “If you don’t use Facebook, you’re behind,” he said. “Even grandmas, everyone was on Facebook.”

To capture customers, Myanmar’s mobile phone operators began offering a sweet deal: use Facebook without paying any data charges.

“Facebook should be sending flowers to me, because we have been an accelerator for bringing the penetration,” said Lars Erik Tellmann, who until July was chief executive of Telenor Myanmar, part of Norway’s Telenor Group. “This was an initiative we took fully on our own. And this was extremely popular.”

In Myanmar today, the government itself uses Facebook to make major announcements, including the resignation of the president in March.


In the fall of 2013, Aela Callan, an Australian documentary filmmaker studying at Stanford University, began a project on hate speech and false reports that had spread online during conflicts between Buddhists and Rohingya Muslims the prior year. In June 2012, at least 80 people had died in riots and thousands of Rohingya were moved into squalid internment camps. Anti-Rohingya diatribes appeared on Facebook. One Buddhist nationalist group set up a page called the “Kalar Beheading Gang.”

In November 2013, she met at Facebook’s California headquarters with Elliott Schrage, vice president of communications and public policy. “I was trying to alert him to the problems,” she said.

Emails between the two show that Schrage put Callan in touch with, a Facebook initiative to bring the internet to developing countries, and with two Facebook officials, including one who worked with civil-society organizations to assist the company in coping with hate speech.

“He didn’t connect me to anyone inside Facebook who could deal with the actual problem,” she said.

Asked for comment, Schrage referred Reuters to a press person at Facebook. The company didn’t comment on the meeting.

Matt Schissler, a doctoral student at the University of Michigan, said that between March and December 2014, he held discussions with Facebook officials in a series of calls and online communications. He told them how the platform was being used to spread hate speech and false rumours in Myanmar, he said, including via fake accounts. He and other activists provided the company with specific examples, including a Facebook page in Burmese that was called, “We will genocide all of the Muslims and feed them to the dogs.” The page was removed.

Schissler belonged to a private Facebook group that was set up so that Myanmar human rights activists, researchers and company employees such as Asia Pacific policy chief Garlick could discuss how to cope with hate speech and other issues. The activists brought up numerous problems with Facebook’s multi-step reporting system for problematic content. As one example, they cited a photograph of an aid worker in Rakhine State in a post that called him “a traitor to the nation.” It had been shared 229 times, according to messages reviewed by Reuters.

One of the private group’s members had reported it to Facebook as harassment of an individual but later received a message back: “We reviewed the photo you reported for containing hate speech or symbols and found it doesn’t violate our Community Standards.” After multiple complaints by activists over six weeks, a Facebook employee finally explained to the activists that the takedown request was rejected because the photo had been reported, but not the comment above it. It eventually was taken down.

In March 2015, Schissler gave a talk at Facebook’s California headquarters about new media, particularly Facebook, and anti-Muslim violence in Myanmar. More than a dozen Facebook employees attended, he said.

Two months later, Madden, the tech entrepreneur, gave a talk at Facebook headquarters about tensions and violence between Buddhists and Muslims. He said he showed a doctored picture that had spread on Facebook of the country’s de facto leader, Aung San Suu Kyi, who is Buddhist, wearing a Muslim head scarf. The image, Madden said, was meant to imply she was sympathetic to Muslims – a “very negative message” in Myanmar.

“The whole point of this presentation was really just to sound the alarm, to show very vividly the context in which Facebook was operating, and already the evidence of how it was being misused,” he said. He left the meeting thinking his audience took the talk seriously and would take action.

Madden had founded a technology hub and start-up accelerator in Yangon called Phandeeyar. He said he and others involved with the venture interacted with Facebook “many dozens” of times over the next several years, including via email, in the private Facebook group and in person, showing how the network’s systems for detecting and removing dangerous content were ineffective. He isn’t sure what steps the company took in response. “The central problem is that the mechanisms that they have to pull down hate speech in a timely way, before it does real world harm, they don’t work,” he said.

Madden and Jes Kaliebe Petersen, Phandeeyar’s chief executive, said Facebook was still relying too much on their group and other volunteers to report dangerous posts. “It shouldn’t be incumbent on an organisation like ours or people who happen to be well-connected with folks inside Facebook to report these things,” Petersen said.

In April, shortly before Zuckerberg’s Senate testimony, Phandeeyar and five other Myanmar groups blasted him for claiming in an interview with Vox that Facebook’s systems had detected and removed incendiary messages in September last year. “We believe your system, in this case, was us,” they wrote. Zuckerberg apologised.

Back in 2014, tech organizations and researchers weren’t the only ones sounding alarms with Facebook. So was the Myanmar government.

In July of that year, riots broke out in the central city of Mandalay after false rumours spread online, on Facebook and elsewhere, that a Muslim man had raped a Buddhist woman. A Buddhist man and a Muslim man were killed in the fighting.

The Myanmar government asked Tun, then a Deloitte consultant, to contact the company. He said he didn’t succeed at first, and the government briefly blocked Facebook.

Tun said he eventually helped to arrange meetings between the government and Facebook. “What they promised to do was, when you spot fake news, you could contact them via email,” Tun said of Facebook. “And they would take action – they were willing to take down pages after their own verification process.”

The government began reporting cases to Facebook, but Tun said he quickly realized the company couldn’t deal with Burmese text. “Honestly, Facebook had no clue about Burmese content. They were totally unprepared,” he said. “We had to translate it into English for them.”


In August 2013, Zuckerberg announced a plan to make the internet available for the first time to billions of people in developing countries.

“Everything Facebook has done has been about giving all people around the world the power to connect,” he said. The company would now work, he added, to make “internet access available to those who cannot currently afford it.”

But in Myanmar, the language barrier would cause trouble. Most people here don’t speak English. Although Myanmar users at the time could post on Facebook in Burmese, the platform’s interface – including its system for reporting problematic posts – was in English.

Making matters worse, the company’s operation for monitoring content in Burmese was meagre.

In 2014, the social media behemoth had just one content reviewer who spoke Burmese: a local contractor in Dublin, according to messages sent by Facebook employees in the private Facebook chat group. A second Burmese speaker began working in early 2015, the messages show.

In Manila – the original site of the outsourced Project Honey Badger – there were no content reviewers who spoke Burmese. People who reviewed Myanmar content there spoke English.

“In cases like hate speech where we didn’t understand the language, we would say, ‘I don’t know the language,’” said a person who worked there. “So the client had to solve that,” the person said, referring to Facebook.

By 2015, Facebook had around four Burmese speakers reviewing Myanmar content in Manila and Dublin. They were stretched thin: that year Facebook had 7.3 million active users in Myanmar.

Accenture slowly began to hire more Burmese speakers. With the help of volunteer translators, Facebook also introduced a Burmese-language interface.

By 2016, the Honey Badger project had moved to Kuala Lumpur after Accenture convinced Facebook it would be easier to recruit Burmese and others to work in Malaysia’s capital than in further-off Manila, according to a person familiar with the matter.

In an office tower in Kuala Lumpur, teams of content monitors are assigned to handle different Asian countries, not just Myanmar. They are paid around $850 to $1000 a month and are often employed by temporary staffing agencies, according to ex-employees and online recruitment ads.

Facebook said in a statement: “We’ve chosen to work only with highly reputable, global partners that take care of their employees, pay them well and provide robust benefits – this includes Accenture in Asia Pacific.”

A spokesperson for Accenture confirmed it partners with Facebook. “The characterization of our operations as ‘secretive’ is misleading and confidentiality is in place primarily to protect the privacy and security of our people and the clients we serve,” the spokesperson said.


Former content monitors said they often each had to make judgements on 1,000 or more potentially problematic content items a day, although the number is now understood to be less.

Facebook’s complete rules about what is and isn’t allowed on its platform are spelled out in its internal community standards enforcement guidelines, which the company made public for the first time in April. It defines hate speech as “violent or dehumanising speech, statements of inferiority, or calls for exclusion or segregation” against people based on their race, ethnicity, religious affiliation and other characteristics.

In response, Facebook said: “Content reviewers aren’t required to evaluate any set number of posts … We encourage reviewers to take the time they need.”

A Facebook official also told Reuters the community standards policy is global, “but there are local nuances,” such as slurs, that content reviewers who are native speakers can consider when making decisions. But former content monitors told Reuters the rules were inconsistent; sometimes they could make exceptions and sometimes they couldn’t.

Former content monitors also said they were trained to err on the side of keeping content on Facebook. “Most of the time, you try to give the user the benefit of the doubt,” said one former Facebook employee.

The ex-monitors said they sometimes had as little as a few seconds to decide if a post constituted hate speech or violated Facebook’s community standards in some other way. They said they didn’t actually search for hate speech themselves; instead, they reviewed a giant queue of posts mostly reported by Facebook users.

Many of the millions of items flagged globally each week – including violent diatribes and lurid sexual imagery – are detected by automated systems, Facebook says. But a company official acknowledged to Reuters that its systems have difficulty interpreting Burmese script because of the way the fonts are often rendered on computer screens, making it difficult to identify racial slurs and other hate speech.

Facebook’s troubles are evident in a new feature that allows users to translate Burmese content into English. Consider a post Reuters found from August of last year.

In Burmese, the post says: “Kill all the kalars that you see in Myanmar; none of them should be left alive.”

Facebook’s translation into English: “I shouldn’t have a rainbow in Myanmar.”

In response, Facebook said: “Our translations team is actively working on new ways to ensure that translations are accurate.” The company said it uses a different system to detect hate speech.

Guy Rosen, vice president of product management, wrote in a blog post on Facebook in May about the problems the company faced in identifying hate speech. “Our technology still doesn’t work that well and so it needs to be checked by our review teams,” he wrote.

Facebook officials say they have no immediate plans to hire any employees in Myanmar itself. But the company does contract with local agencies for tasks unrelated to content monitoring. One is Echo Myanmar, a communications firm whose managing director is Anthony Larmon, an American.

Larmon has expressed strong opinions on the Rohingya. Toward the end of 2016, the Myanmar army launched an onslaught across some 10 villages after Rohingya militants attacked border posts. At the time, a U.N. official accused the government of seeking “ethnic cleansing” of the Rohingya.

In November 2016, Larmon wrote that an article about the U.N. allegation was “misleading.” He cited what he said were claims by multiple “local journalists” that the ethnic minority “purposely exaggerate (lie about)” their situation to “get more foreign aid and attention.”

He also wrote: “No, they aren’t facing ethnic cleansing or anything remotely close to what that incendiary term suggests.” He said he later removed the post.

A Facebook spokesperson said that Larmon’s post “does not represent Facebook’s view.”

Larmon told Reuters: “It was overly-emotional, under-informed commentary on a highly nuanced subject that I do regret. My view on the Rohingya, same today as then, is that they should be safely repatriated and protected.”

The platform on which he aired his views about the Rohingya? Facebook.



Source link

Continue Reading