TSPA’s advisors are people who work within and adjacent to the online trust and safety field. They provide us with deeper insight into current and emerging trends and concerns within the field, and ensure that we incorporate broad and diverse perspectives into our work. They provide expert information on particular subjects, and often work alongside our members within our working groups and on special initiatives.
Christine Chen has worked at the intersection of technology, policy, communications, and international affairs for more than two decades. She formerly led international and content policy communications at Facebook. Previously, she worked at Google, where her roles included leading public policy at YouTube, managing public policy strategy around free expression and international relations at Google, and leading privacy communications. Christine previously served on the board of directors at the Global Network Initiative. Prior to her career in Silicon Valley, she was a magazine journalist at Foreign Policy, Fortune, and Newsweek.
Dali Szostak is head of UX for Google Trust and Safety leading a team of designers, researchers and content strategists with the goal of shaping the user experience of all the products, systems and tools that counter abuse in Google platforms. She shaped her knowledge of digital security by speaking to (and being inspired by) countless activists, human rights defenders, journalists and experts during pivotal times in places like Ukraine, Zimbabwe, Kenya, Venezuela, and India, leading Google in the understanding of the people using our products that are at high risk. One such group of people are Content Moderators, individuals who are at the front lines of keeping our digital lives safe and are not immediately considered as key stakeholders in the way that technology is designed. Besides bringing a product and community design lens to Trust and Safety, she is also an educator, in particular supporting the growth of UX and digital safety in Latin America.
Daphne Keller is the Director of Platform Regulation at Stanford University’s Cyber Policy Center, where her work focuses on platform regulation and Internet users’ rights. She has published both academically and in mainstream media; testified and participated in legislative processes; and taught and lectured extensively. Her recent work focuses on legal protections for users’ free expression rights when state and private power intersect. Until 2015, Daphne was Associate General Counsel for Google, where she had primary responsibility for the company’s search products. She worked on groundbreaking intermediary Liability litigation and legislation around the world and counseled both overall product development and individual content takedown decisions.
Dona Bellow currently leads Policy & Responsible Innovation programs at IRL, a group messaging startup. She began her tech career at Google, where she led the policy evaluation and operationalization for legal content removal requests in the French market and managed the Child Safety program for Legal Online Operations, where she also became very involved in reviewer wellness and mental health. She went on to co-develop a program supporting product teams in mitigating abuse-related risks. She then joined the Community Policy team at Airbnb, collaborating closely with public policy teams on regulatory risks. Later, she worked on creating processes for trust and safety policy impact at Twitter; and most recently was a Responsible Innovation Manager at Meta. Dona is a proud native of Togo, and grew up in France.
Eric Davis has safeguarded brands, platforms, and billions of users from bad ads, bad apps, and other machinations of bad actors across global policy, product, and engineering functions throughout his career. He was Senior Director for Product Management at NortonLifelock (previously Symantec), where his work included leading product requirements and implementation addressing GDPR and other privacy regulations globally. Previously, he worked for 13 years at Google, where he held product policy and public policy leadership roles. Eric also founded and led Google’s Android Security Investigations and Operations, Anti-Malvertising, and Trust and Safety teams. Earlier in his career, he was the charter International Product Manager for Trust & Safety at eBay.
Fatima Alam is a Global Compliance Manager at Netflix, where she works on content classification across a Global Content Regulation and Standards Board portfolio. Fatima previously worked in various roles on the Trust and Safety and Public Policy teams at Google in the US and India, where she supported and led the research, development, and enforcement of product policies related to controversial and harmful online content. She is also interested in the ways policy decisions about science and technology are made by governments around the world and the various roles played therein by technology platforms, elected representatives, bureaucrats, subject matter experts, and civil society. Previously, she was also a Human Rights and Technology Fellow at the Harvard Carr Center.
Jess Miers is a recent graduate from Santa Clara University School of Law where she studied Internet law and technology policy. Her scholarship primarily covers Section 230 and content moderation. During law school, Jess was a legal intern for Twitter, TechFreedom, and the UCLA Institute for Technology Law and Policy. She also founded the SCU Internet Law Student Organization with the goal of inspiring the next generation of Internet advocates and content specialists. Jess is presently employed at Google as a Government Affairs & Public Policy Analyst where she works with a team of subject matter experts on an array of issues pertaining to access to information and intermediary liability. Before joining Public Policy, Jess was a product lead for Google Trust & Safety where she worked extensively on legal removals, operations, and content policy enforcement in the U.S. and E.U.
Jerrel Petersonhas spent his career leveraging research and strategic partnerships to address complex social and political issues. He is currently a Senior Policy Manager on the Trust & Safety team at Spotify and previously served as the Head of Safety Policy for Twitter. He worked cross functionally to resolve dozens of high-profile cases weekly, drive the development of policies, processes, and tools, and communicate Twitter’s content policies clearly and objectively to internal and external stakeholders. Prior to Twitter, Jerrel worked in public policy at the state and federal level and also spent a few years providing direct services to various marginalized communities within the U.S.
Kaitlin Sullivan is a Director of Content Policy at Facebook, where she leads an international team that develops and scales policies to address account-level and behavioral abuse. She partners with Product and Operations teams to set company strategy for effective, transparent, and principled content moderation. Kaitlin regularly speaks on these issues at universities and conferences, including the Conference on Crimes Against Women, Content Moderation at Scale, and the Internet Governance Forum. Prior to joining Facebook, Kaitlin worked as a rape crisis advocate and educator at the YWCA of Silicon Valley.
Nicole Wong specializes in assisting high-growth technology companies to develop international privacy, content, and regulatory strategies. She previously served as Deputy U.S. Chief Technology Officer in the Obama Administration, where she focused on internet, privacy, and innovation policy. Prior to her time in government, Nicole was Google’s Vice President and Deputy General Counsel, and Twitter’s Legal Director for Products. She frequently speaks on issues related to law and technology, including five appearances before the U.S. Congress. Nicole chairs the board of Friends of Global Voices, and also sits on the boards of WITNESS, the Mozilla Foundation, and The Markup. Nicole currently serves as co-chair of the Digital Freedom Forum, and as an advisor to the AI Now Institute, the Alliance for Securing Democracy, Luminate, Refactor Capital, and the Albright Stonebridge Group.
Pinal Shah’s work in tech policy centers on two main themes: the behavioral impacts of technology on individuals and society, and racial and social justice. Pinal is currently a Sr. Behavioral Engineer at Robinhood, where she is building out Trust & Safety functions. Her work focuses on countering misinformation & disinformation, content moderation, and Safety by Design to ensure transparency and safety are built into products & features so as to avoid unintended consequences. Prior to joining Robinhood, Pinal helped build the Privacy team at Lyft, crafting data ethics practices, safeguarding users’ private data, and ensuring compliance with CCPA and GDPR. Under the Obama Administration, she was appointed to advise at the Department of Homeland Security’s Science & Technology Directorate, working on policies and programs to safeguard the nation, including developing partnerships with foreign partners for information and technology sharing, Unmanned Aerial Vehicle policy, and the SAFETY Act. She is currently a member of the “Virtually Human” working group at Aspen Digital, and is a proud alum of Howard University School of Law.
Treesa Ann Jose
Treesa Ann Jose has worked at the intersection of technology and policy for close to a decade. She currently works as a Product and Content Policy Manager at Facebook. Previously, she worked at Google where she set up and scaled Operations teams and Content Policy strategy for the Next Billion User offerings. Her Policy and Product expertise spans across User Generated Content, Elections Integrity, Business Content including integrity strategy for Augmented and Virtual Reality applications. Beyond her core Tech Policy focus, Treesa has worked with international and regional NGOs like World Economic Forum on community development and women empowerment efforts throughout her career in Singapore, India and the United States.
Alex Feerst | Former GC and Head of T&S, Medium.com
Tim Lordan | Executive Director of Internet Education Foundation (IEF)
Micah Shaffer | Former Director of Public Policy, Snap
Role and Expectations of Advisors
TSPA’s advisors inform and guide our decision-making on various programming, activities, and relevant current issues. Advisors’ responsibilities are to: (1) be a trusted source of insight and information to the trust and safety field, (2) identify and connect TSPA to experts who can offer advice on organizational matters, (3) expand TSPA board and staff’s network with other communities, organizations, and individuals, and (4) boost the work of TSPA in the industry and beyond.
Advisors are expected to be active participants. At minimum, advisors are requested to:
- Commit to at least a 1-year term
- Participate in two advisors meetings a year
- Provide advice or support to the organization throughout the year
- Support and participate in TSPA programming and activities
- Keep confidential non-public information shared by TSPA, TSF, or fellow advisors
Advisors will not be compensated for their time or advice, but may be hired as consultants for specific projects. Advisors do not have the authority to vote on organizational matters unless specifically requested by TSPA staff or boards. Advisors are not representatives of, nor do they speak on behalf of, their current employers. Advisors do not need to be members of TSPA.
How to Become An Advisor
In June of each year, we will refresh our group of advisors. This refresh will include (1) identifying and re-inviting current advisors to serve another term, (2) identifying specific gaps within the advisors group and asking qualified candidates to apply, and (3) opening the application process to the broader public.
Process for New Advisors
- Individuals interested in becoming an advisor should fill out an application.
- TSPA staff may get in touch with applicants if more information is required.
- TSPA staff will review and recommend selected applicants.
- TSPA staff, board members, and current advisors will vote on the recommended applicants. Applicants with the most votes will be invited to become an advisor. Number of applicants accepted will be based on the number of open spots.
Process for Nominated Advisors
Throughout the year, TSPA staff, board, and current advisors may nominate a candidate to join as an advisor to meet a gap in expertise or perspective. Since these candidates are specifically nominated to meet an identified need, they will not be formally voted on by the full staff and board; however, informal agreement by a majority of the staff and board will be required.
Process for Advisors Invited to Rejoin
TSPA staff may invite advisors to rejoin after their terms have expired. This invitation may be decided based on an advisor’s level of participation, their involvement in current or planned programming or activities, or because their perspective and knowledge is deemed essential. Current advisors who are asked to rejoin will be decided by staff and will not require voting or agreement.
TSPA advisors must have (1) deep knowledge of and experience with trust and safety issues online, and/or (2) expertise in one or more of the following areas as they relate to online trust and safety: content/behavior policy, operations (including content review), internet regulation (US and international), human rights, product development, public policy and communications, labor relations, or people management.
Additional prioritization criteria include:
- Candidates who have specific expertise or skills that will benefit current or planned programming or activity and organizational development.
- Candidates who reflect the diversity of our membership in regards to:
- The various types of roles within trust and safety work
- Experience at employers or organizations of different sizes and sectors
- Geographical location
- Representing different perspectives and backgrounds
Term of Service
Advisors will serve starting from when they have been selected either through the selection process (generally July 1) or through the nomination process (throughout the year). All advisors’ terms of service end June 30 unless invited by TSPA to rejoin for the next year.
Advisors may be asked to withdraw as an advisor if they develop a conflict of interest, as determined by TSPA staff and boards. An advisor can also request to resign at any time.
Apply to Become an Advisor
We are not currently accepting applications for the 2021-2022 period. Advisors applications will re-open in June 2022.