This is an HTML version of an attachment to the Official Information request 'Correspondence with TikTok regarding Christchurch call'.

1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 

The information contained in this email message is for the attention of the intended recipient only and is not necessarily the official view 
or communication of the Department of the Prime Minister and Cabinet. If you are not the intended recipient you must not disclose, 
copy or distribute this message or the information in it. If you have received this message in error, please destroy the email and notify 
the sender immediately.  
1982
Act 
Information 
Official 
the 
under 
Released 
3


1982
Act 
Information 
Official 
the 
under 
Released 

s6(a)
1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


 
 
 
While our expert teams pay attention to the nuance of our enforcement, we 
also run databases of keywords associated with violent extremism and hate, 
which disallow users from searching for particular terms. Similarly, we maintain 
databases of signs, symbols and images related to violent extremism, 
dangerous organisations and hateful ideologies to ensure content containing it 
is disallowed, blocked from upload or removed. TikTok continuously reviews 
entities that are know to participate in hateful behaviour, violent extremism ,  1982
and criminal activities. These identified groups are not able to join our platform 
and action would be taken against any such accounts. Additionally, TikTok has 
commercial partners that monitor and gather intelligence on violent extremist 
Act 
content and provide threat alerts on credible threats. We have access to 
databases that  include more than 7,000 archived audio/video files and 200,000 
URLs published/shared by designated extremist groups. We use this 
intelligence to improve our detection and proactive risk mitigation. 
 
These are some of the ways that TikTok is already deploying technical solutions 
and product changes that make our platform experience a safer one for users. 
We will continue to innovate in this space, and look forward to working with 
Information 
the Coalition to identify opportunities to enhance protections.   
 
We know there is no finish line when it comes to further improving safety on 
TikTok and our transparency reports show that we removed 2,712,492 pieces 
of content that violated our Community Guidelines relating to violent extremism 
Official 
in 2021.  
 
2. Referring to the Call to Action, what commitments has TikTok 
the 
already addressed or taken measures to address?  
 
Through TikTok's publicly available Community Guidelines and Terms of 
Service, our platform clearly articulates and enforces policies to support the Cal  
to Action commitments.  
under 
 
Violent extremism, dangerous organisations and hateful ideologies are explicitly 
designated in our Community Guidelines as being prohibited on our platform. 
And our Community Guidelines Enforcement Reports are regularly updated to 
provide information on the types of content we remove, at what volume, and 
what stage of moderation. Our most recent update is available here. 
 
Released 
As described in the answer above, we use a mixture of machine and human 
moderation to detect, remove and take enforcement action as appropriate 
when violative content is detected on TikTok. Specifically, TikTok has a clear 
policy of removing violent extremism content. We reject any form of violent 


 
content or content that seeks to glorify groups and/or individuals associated 
with violent extremism from using our platform to advance their violent 
extremism propaganda.  
 
We will terminate accounts and may also ban users from coming back to our 
platform if they have posted content that glorifies and promotes violence.  
 
While we take decisive action against violative content and accounts, we also 
ensure we notify our users of the actions taken and provide an avenue to  1982
appeal, consistent with transparent, best practice.  
 
In late 2021, we provided an update on work we are doing regarding our 
Act 
recommendation system, to keep people's "For You Feed" fresh and to interrupt 
repetitive patterns to foster diversity of content. Further details are available 
here, and it builds on our previously provided overview of how our 
recommendation system works, which is available here.  
 
TikTok has a network of Transparency and Accountability Centres, that provide 
outside experts an opportunity to directly view how our teams at TikTok go 
about the day-to-day challenging, but critically important, work of moderating 
Information 
content on the platform, our privacy policies and other important practices. We 
would be pleased to welcome representatives of the Christchurch Cal  an 
opportunity to participate in a tour of our Transparency and Accountability 
Centre. 
 
Official 
TikTok also works with industry experts, non-governmental organisations, and 
industry associations throughout our safety policy and feature development 
process to monitor emerging trends, identify blindspots or gaps in our 
the 
knowledge, and generate innovative solutions for building the safest possible 
platform for our community.  
 
It's worth noting that TikTok currently participates in three of the Global 
Internet Forum to Counter Terrorism's (GIFCT) working groups, in order to 
under 
learn from and contribute to industry best practices. These working groups are: 
Transparency, Crisis Response, and Positive Interventions and Strategic 
Communications. TikTok continues to actively work toward gaining 
membership in GIFCT and recently completed the GIFCT-Tech Against 
Terrorism mentorship program. 
 
3. How does TikTok intend to contribute to and participate in the 
Released 
Christchurch Call Community?  
 
 
TikTok's top priority is safety for our community. We know there is no finish 
line when it comes to keeping our users safe and we are committed to working 


 
col aboratively with our industry peers to ensure we are playing a proactive and 
meaningful role globally to support shared safety goals. 
 
TikTok is committed to sharing information that will support the identification 
and suppression of terrorist and violent extremist content. We do so already 
through global and regional initiatives, including: 
x  MHA-RSIS Webinar on "Countering Extremist Content: Promoting 
Positive Narratives Online" - an engagement to combat extremism online  1982
with the Singapore Government's Home Affairs Ministry and a local 
think-tank.  
x
Act 
  Collaboration with UNDP to empower content creators in APAC to build 
a safe online space  
Our intention is to continue this collaborative and open engagement through 
the Christchurch Call community and we are open to more detailed discussions 
around how we can most meaningfully support achieving the Christchurch Call's 
objectives. 
 
4. Could you please describe TikTok’s ownership structure? What is 
Information 
the nature of the relationship between ByteDance and Beijing 
ByteDance Technology (noting the Chinese Government’s recent 
acquisition of a 1% stake in the latter)? a. Does this investment allow 
for a board seat in Beijing ByteDance Technology? How does this 
relate to TikTok’s operations?
 
 
Official 
TikTok Ltd is owned by Bytedance Ltd, a privately owned technology company 
incorporated in the Cayman Islands, and the majority investors include 
the 
Bytdance and TikTok founder, Yiming Zhang and large institutional investors 
such as Sequoia Capital, General Atlantic, KKR and Softbank. We have a global 
board of directors.  
 
TikTok’s CEO is based in Singapore, and we have a diverse leadership team, 
under 
many of which are based in the United States. The TikTok app is available in 
more than 150 countries, but not in China.  
 
The Chinese Government has not taken an ownership stake in TikTok.  
 
The subsidiary referenced has no ownership, visibility, or input into TikTok’s 
operations. 
Released 
 
There are no TikTok companies incorporated in China.  
 


 
In summary, there is nothing about TikTok's ownership structure which would 
limit our commitment to participate in the Christchurch Call. 
 
5. What assurances can TikTok give about its data protections and the 
independence of its content moderation processes? 
 
New Zealand user data is stored in the US and Singapore. As indeed is the data 
of all TikTok users, worldwide.  
 
1982
We have strict access controls in place - including restrictions on who has 
access to which data set, strong authentication measures, logging of access, 
limited access period, and encryption of data. We employ thousands of safety 
Act 
professionals around the world to ensure we have 24/7 coverage of content on 
TikTok. 
 
Our moderation teams that review New Zealand content (for example) are 
based in Europe and SE Asia and provide coverage on a 24/7 basis. This team 
is supervised by dedicated experts in policy and enforcement who write the 
content moderation guidelines and leads practice - they are based in our 
regional Trust & Safety hubs in Singapore, Dublin, and California and include 
Information 
specialists who are familiar with our local culture and norms.  
 
We moderate content according to our Community Guidelines and local laws, 
our moderation teams are empowered to make decisions that suit the local 
market. 
Official 
 
6. Noting ByteDance products in China, such as Douyin, are subject 
to regulatory frameworks and privacy regimes there, to what extent 

the 
is TikTok also subject to these and/or, as a subsidiary of ByteDance, 
has the potential to be subject to these?
 
 
TikTok is not available in China and we are a global company. We have not and 
will not share user data with the Chinese Government, and would not do so if 
under 
asked.  
 
Any legal requests by any Government for TikTok user data would be sent to 
our Law Enforcement Response Team (LERT) and would follow a well 
established international process. Such requests would need to go through the 
relevant Mutual Legal Assistance Treaty (MLAT) process. 
 
Released 
As is made publicly available in our regularly published Transparency and 
Accountability reports, the Chinese Government has not requested or asked for 
TikTok user data. We have never given the Chinese Government TikTok user 
data. 


 
 
Our Trust and Safety Centres are located in Singapore, Ireland and the US; our 
data storage facilities are in US and Singapore (which have achieved 
international recognised security certifications), and we have strict controls on 
who can access it. 
 
We also have a dedicated US based Data Protection team focused on controls 
implementation and monitoring of our protected sensitive data. and our Chief 
Security Officer, Roland Cloutier, brings to TikTok over 30 years of experience  1982
in cyber security and law enforcement, including over a decade serving the 
United States Air Force, Department of Defence and Department of Veterans 
Affairs.  
Act 
 
We would welcome conveners of the Christchurch Call to participate in a 
Transparency and Accountability Centre tour which will provide you an 
opportunity to visit and see first-hand how teams at TikTok go about the 
critically important work of securing our community's safety, data, and privacy. 
 
7. How does TikTok approach the concept of a free, open and secure 
internet? Does TikTok have policies and/oral public statement on how 

Information 
it respects and protects human rights and fundamental freedoms? 
 
TikTok has progressed a range of partnerships and projects related to human 
rights, including as a member of Business for Social Responsibility (BSR) and 
TikTok has joined BSR's Human Rights Working Group. Together with BSR and 
Official 
Article 1 Advisors, we have held educational trainings for senior leadership on 
the UN Guiding Principles on Business and Human Rights, and are committed 
to respect human rights under the UNGP's Protect-Respect-Remedy framework.  
the 
 
We have also publicly made a statement with respect to our approach to human 
rights that is available on our website. It reads: 
 
Technology is an essential gateway to the exercise of human rights. TikTok 
under 
touches more than 1 billion people across the globe, who represent a variety 
of ethnicities, orientations, backgrounds and life experiences. As a global 
entertainment company, TikTok has a responsibility to ensure that our 
community is treated. 
 
Responsibility for upholding human rights is shared: while governments have 
the responsibility to protect human rights, TikTok and other businesses have a 
Released 
responsibility to respect those human rights. Respecting human rights is 
essential for TikTok to build and sustain trust among our employees, creators, 
advertisers, and others who engage with our company. 
 


 
Our philosophy is informed by the International Bill of Human Rights (which 
includes the Universal Declaration of Human Rights and the International 
Labour Organisation’s Declaration on Fundamental Principles and Rights at 
Work) and the United Nations Guiding Principles on Business and Human 
Rights. As part of our commitment, we will strive to respect human rights 
throughout our business and will comply with applicable laws and regulations 
intended to promote human rights where we conduct business globally. We will 
continuously evaluate our operations to identify, assess, and address salient 
human rights risks; engage key stakeholders; and prioritise key areas where  1982
we have the greatest opportunity to have a positive impact. 
 
8. What is TikTok doing to improve transparency of the platform?  
Act 
 
As articulated in our answer to Question 2 above, at TikTok we recognise that 
accountability and transparency are essential to facilitating trust, and we are 
committed to being transparent in how we operate, moderate and recommend 
content, and secure our platform. In early 2020 TikTok launched its first 
Transparency Report. As part of that work, we provide regular updates on 
reports received across different violation types, the frequency of those reports, 
and how much of that content we act on for violating our Community 
Information 
Guidelines. We also disclose the percentages of content we take action on 
proactively, through automation and detection models that enqueue content 
for human moderation. 
 
As mentioned, TikTok has also opened Transparency and Accountability 
Official 
Centres for experts and policymakers to see first-hand how we are working to 
build a safe and secure platform for our growing and diverse community. Due 
to physical constraints around the ongoing COVID-19 pandemic, opening the 
the 
centre has meant evolving our original plans to offer a virtual experience until 
we can welcome guests to the full tour at our physical locations.  
 
Moving forward, we endeavour to lead the industry in transparency including 
through robust transparency reporting as well as additional experiences at the 
under 
Transparency and Accountability Centres, including opportunities for qualified 
experts to review our algorithm. 
 
We are also very upfront about our platform's policies, including our Terms of 
Service and our Community Guidelines. Everyone who uses TikTok is bound by 
our Terms of Service, and our Community Guidelines apply universally. We take 
a firm stance on behaviours and content which are not in keeping with our 
Released 
policies, and we will not hesitate to take action against accounts that might 
seek to undermine the safety of our community.  
 


 
When users make reports about particular content, we notify them of the 
outcome of that report. For TikTok users whose content has been removed or 
account has terminated, we notify them with information on why the action 
was taken, and they will be allowed  to submit an appeal should they feel the 
removal decision was unjustified.  
  
9. How does TikTok moderate live content for TVEC?  
 
At TikTok, we use a number of measures to identify and moderate TVEC,  1982
including through automated detection. 
 
Specifically, we have multiple models that detect content that overlaps with 
Act 
Violent Extremism (VE) signals, for example (but not limited to), we have 
model(s) that detects blood in the frame and another that detects sensitive 
military or police clothing. If our models are hit with a positive match, it will be 
automatically sent to the moderation queue where human moderators will 
assess the content for possible violations. 
 
We also have a dedicated high risk detection database, which scans through 
high risk content (such as footage from previous Violent Extremism attacks) 
Information 
that will automatically detects and remove this content. 
 
Regarding our moderation practices, our moderators are trained on integrated 
policies including being able to detect violent extremist signals and enforce 
against them. 
Official 
 
We also have a number of dedicated teams that help moderate live content for 
TVEC including: 
the 
x  a dedicated Emergency Response Team that work with law-enforcement 
to tackle the use of VE actors on the platform;  
x  an Incident Management Team that handles major VE events from an 
operational perspective and collaborates closely with our policy team;  
under 
x  a Risk Analysis Team that proactively detects possible threats to the 
platform for teams to prepare against.  
Our global policy team has global and regional experts focusing on violent 
extremism, hateful organisations and ideologies, including policy experts 
dedicated to livestreams. 
 
Released 
Our Trust and  Safety teams pay special attention to livestreams, taking into 
consideration the unique nature of this feature, including but not limited to the 
simultaneity between creation and consumption of content, the interaction 
between viewers and hosts, existing or past strikes against user accounts.  Our 


 
policies and safety strategies are continuously reviewed and updated: they are 
influenced and shaped by real-life events and feedback from experts. As violent 
extremist events occur around the world, we work to proactively draft policies 
that are able to scale and adapt to novel concepts. 
 
10. How does TikTok prevent recommending TVEC or harmful 
content? How does TikTok assess and moderate algorithmic 
recommendation processes? 
 
 
1982
As previously described, we use a mix of technology and human moderation to 
enforce our Community Guidelines (which includes controlling TVEC or harmful 
content), which is also informed by in-app user reporting (and non-users via 
Act 
website), where users have an option to easily report potentially violative 
content.  
 
All content uploaded to TikTok passes through technology that works to identify 
potential policy violations. Our safety team takes additional precautions to 
review videos as they rise in popularity to reduce the likelihood of content that 
may not be appropriate for a general audience entering our recommendation 
system. 
Information 
 
With regard to our recommendation system, we present users with a stream of 
videos on our ‘For You’ feed curated to their interests, but also works to 
intersperse recommendations that might fall outside people’s expressed 
preferences, offering an opportunity to discover new categories of content.  
Official 
 
Our recommendation system on TikTok is designed with safety as a key 
consideration. Reviewed content found to depict things like graphic medical 
the 
procedures or legal consumption of regulated goods, for example - which may 
be shocking if surfaced as a recommended video to a general audience that 
hasn’t opted in to such content - may not be eligible for recommendation. Users 
can also report a video that seems out of line with our guidelines. All these 
actions contribute to future recommendations on the For You feed.  
under 
 
Content along the lines of TVEC or harmful content in general will not be eligible 
for recommendation under any circumstances and wil  be taken down from the 
platform as they violate our Community Guidelines.  
 
Ultimately, the For You feed is powered by user’s feedback: the system is 
designed to continuously improve, correct, and learn from user’s engagement 
Released 
with the platform to produce personalised recommendations that we hope 
inspire creativity and bring joy with every refresh of their For You feed.  


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released 

 
The information contained in this email message is for the attention of the intended recipient only and is not necessarily the official view 
or communication of the Department of the Prime Minister and Cabinet. If you are not the intended recipient you must not disclose, 
copy or distribute this message or the information in it. If you have received this message in error, please destroy the email and notify 
the sender immediately.  
  
 
The information contained in this email message is for the attention of the intended recipient only and is not necessarily the official view 
or communication of the Department of the Prime Minister and Cabinet. If you are not the intended recipient you must not disclose, 
copy or distribute this message or the information in it. If you have received this message in error, please destroy the email and notify 
the sender immediately.  
1982
  
Act 
Information 
Official 
the 
under 
Released 
3


1982
Act 
Information 
Official 
the 
under 
Released 


1982
Act 
Information 
Official 
the 
under 
Released