Tag: Surveillance

  • Pros and Cons of California SB-1047: The AI Regulation Debate

    Pros and Cons of California SB-1047: The AI Regulation Debate

    Background

    With the recent emergence of ChatGPT, artificial intelligence (AI) has transformed from an obscure mechanism to a widely-used tool in day-to-day life. Around 77% of devices integrate some form of AI in voice assistants, smart speakers, chatbots, or customized recommendations. Still, while at least half of Americans are aware of AI’s presence in their daily lives, many are unable to pinpoint how exactly it is used. For some, the rapid growth of AI has created skepticism and concern. Between 2021 and 2023, the proportion of Americans who expressed concern about AI increased from 37% to 52%. By 2023, only 10% of Americans were more excited than concerned about AI applications in their day-to-day lives. Today, legislators at the federal and state level are grappling with the benefits and drawbacks of regulating AI use and development. 

    California’s SB-1047: An Introduction

    One of the key players in AI development is the state of California, which houses 35 of the 50 most prominent AI companies in the world. Two cities in California, San Francisco and San Jose, account for 25% of all AI patents, conference papers, and companies worldwide. California has responded to the growing debate on AI use through legislative and governmental channels. In 2023, Governor Gavin Newsom signed an executive order establishing initiatives to study the benefits and drawbacks of the AI industry, train government employees on AI skills, and work with legislators to adapt policies for responsible AI development. 

    One such policy that gained attention is SB-1047, or the Safe and Secure Innovation for Frontier Artificial Intelligence Models Act. The bill passed both chambers of the state legislature, but was vetoed by Governor Newsom in September 2024. Introduced by state senator Scott Weiner of San Francisco, SB-1047 aimed to establish safeguards in the development of large-scale AI models. Specifically, the bill applied to cutting-edge AI models that use a high level of computing power or cost more than $100 million to train. Its key provisions included:

    • Cybersecurity protections: Requires developers to take reasonable cybersecurity precautions to prevent unauthorized access to or unintended use of the AI model
    • Pre-release assessment: Requires developers to thoroughly test their AI model for potential critical harm before publicly releasing it. Establishes an annual third-party audit for all developers
    • “Kill switch”: Requires developers to create a capacity to “promptly enact a full shutdown” of the AI program in the case it risks damage to critical infrastructure
    • Safety protocol: Requires developers to create a written safety and security protocol, assign a senior professional to implement it, publish a redacted version, and send an unredacted version to the U.S. Attorney General upon request
    • Whistleblower protections: Prohibits developers from retaliating against employees who report violations of safety protocol internally or to government officials
    • CalCompute: Establishes a publicly-owned and -operated cloud computing infrastructure to “expand access to computational resources” for researchers and startups

    Pros of SB-1047

    One of the main arguments in favor of SB-1047 was that the bill encouraged responsible innovation. Proponents of the bill emphasized that it aligned with federal policy in targeting large-scale systems with considerable computing power, which pose the highest risk of harm due to their cutting-edge nature. They argued that the bill’s holistic approach to regulation, including preventative standards like independent audits and response protocol like the “kill switch” provision, make it difficult for developers to simply check a box stating they do not condone illegal use of their AI model. 

    Proponents also applauded the bill’s protections for whistleblowers at companies that develop advanced AI models. Given the lack of laws on AI development, general whistleblower protections that safeguard the reporting of illegal acts leave a gap of vulnerability for AI workers whose products are largely unregulated. Supporters say SB-1047 would have filled this gap by allowing employees to report potentially dangerous AI models directly to government officials without retaliation. In September 2024, over 100 current and former employees of major AI companies – many of which publicly advocated against the bill – sent a letter to Governor Newsom in support of the legislation’s protections. 

    Other supporters were enthusiastic about the bill’s establishment of CalCompute, a cloud computing infrastructure completely owned and operated by the public sector. Advocacy group Economic Security California praised CalCompute as a necessary intervention to disrupt the dominance of a “handful of corporate actors” in the AI sector. Other advocates emphasized that CalCompute would complement, rather than replace, corporations in providing supercomputing infrastructure. They argued that the initiative would expand access to AI innovation and encourage AI development for public good. 

    Another key argument in favor of SB-1047 is that the bill would have created a necessary blueprint for AI regulation, inspiring other states and even the federal government to implement similar protections. By signing the bill into law, proponents argue, California would have become the “first jurisdiction with a comprehensive framework for governing advanced AI systems”. Countries around the world, including Brazil, Chile, and Canada, are looking at bills like SB-1047 to find ways to regulate AI innovation as its applications continue to expand. 

    Cons of SB-1047

    SB-1047 received criticism from multiple angles. While some labeled the bill an unnecessary roadblock to innovation, others argued for even stronger regulations.

    On one hand, the bill’s large scope was criticized for focusing too heavily on theoretical dangers of AI, hindering innovation that might lead to beneficial advancements. Opponents contended that some of the language in the bill introduced hypothetical scenarios, such as the creation and use of weapons of mass destruction by AI, with no regard to their low plausibility. Major companies like Google, Meta, and OpenAI voiced opposition to the bill, warning that the heavy regulations would stifle productivity and push engineers to leave the state. 

    Others criticized the bill for its potential impacts on academia and smaller startups. Fei-Fei Li, co-director of Stanford University’s Human-Centered AI Institute, argued that the regulations would put a damper on academic and public-sector AI research. Li also stated that the bill would “shackle open source development” by reducing the amount of publicly available code for new entrepreneurs to build off of – a fear that was echoed by national lawmaker Nancy Pelosi (D-CA).

    On the other hand, some believe the bill did not go far enough in regulating cutting-edge AI. These critics pointed to provisions that exempt developers from liability if certain protocols are followed, which raised questions for them about the bill’s ability to hold developers accountable. They also criticized amendments that reduced or completely eliminated certain enforcement mechanisms such as criminal liability for perjury, stating such changes catered to the interests of large tech corporations. Critics argued that the bill’s vague definitions of “unreasonable risk” and “critical harm” leave ample room for developers to evade accountability. 

    Given the bill’s sweeping language in key areas, critics worried that it could either overregulate, or fail to regulate, AI effectively.

    Recent Developments

    On February 27th, 2025, SB-1047 sponsor Scott Weiner introduced a new piece of legislation on AI safety. The new bill, SB-53, was created with a similar intention of safeguarding AI development, but focuses specifically on the whistleblower protection and CalCompute provisions of the original bill.  

    While California continues to grapple with state-level regulations, the federal government has also taken steps to address AI. The Federal Communications Commission is using the 1980s Telephone Consumer Protection Act to restrict AI-generated human voices. The Federal Trade Commission has warned against AI misuse, including discrimination, false claims, and using AI without understanding its risks. In 2024, the Office of Management and Budget issued AI guidelines for all federal agencies. Later that year, the White House formed an AI Council and the AI and Technology Talent Task Force. Although no federal legislation has been passed, these actions show a growing focus on AI regulation.

    Conclusion 

    California’s Safe and Secure Innovation for Frontier Artificial Intelligence Models Act aimed to regulate AI development through novel safeguards. While it was applauded by some as a necessary response to an ever-evolving technology, others believed its wide regulations would have stifled innovation and entrepreneurship. As AI’s use and applications continue to evolve, new policy solutions are likely to emerge at both a state and federal level in the future. 

  • Tulsi Gabbard Confirmed as Director of National Intelligence: Can She Keep America Safe?

    Tulsi Gabbard Confirmed as Director of National Intelligence: Can She Keep America Safe?

    In another significant development in Washington, Tulsi Gabbard—once a Democratic congresswoman, later an outspoken critic of her party, and now a key ally of President Donald Trump—has been confirmed as the Director of National Intelligence (DNI).

    Gabbard’s political journey has been anything but conventional. Born in American Samoa and raised in Hawaii, she became the youngest person elected to the Hawaii state legislature at 21. After serving in the Hawaii National Guard and deploying to Iraq, she was elected to the U.S. House of Representatives in 2012, becoming the first American Samoan and Hindu member of Congress.

    Her tenure in Congress was marked by independent positions on foreign policy, including a 2017 meeting with Syrian President Bashar al-Assad, which drew bipartisan criticism. In 2022, she announced her departure from the Democratic Party, claiming it was under the control of an “elitist cabal of warmongers.” 

    A Contentious Confirmation Process

    During her confirmation hearings, Gabbard faced intense scrutiny over her past statements and actions. Senators questioned her previous defense of Edward Snowden, whom she had once called a “brave whistleblower.” When pressed to label Snowden a traitor, she responded: “Edward Snowden broke the law.” 

    Snowden, a former NSA contractor, leaked classified information in 2013 about the U.S. government’s mass surveillance programs, including the bulk collection of American phone records under the Patriot Act. His revelations exposed the extent of the NSA’s global surveillance operations and sparked a worldwide debate on privacy, national security, and government overreach. While some view Snowden as a whistleblower who revealed unconstitutional surveillance, others, including U.S. officials, see him as a criminal who endangered national security.

    Gabbard had previously argued that Snowden deserved a fair trial rather than immediate prosecution under the Espionage Act, which does not allow whistleblowers to defend their actions in court. However, her confirmation hearing remarks signaled a shift, suggesting she would take a harder stance on intelligence leaks now that she leads the nation’s intelligence apparatus.

    Reversal on Surveillance Policies

    During her confirmation hearings, Tulsi Gabbard faced intense scrutiny over her past statements and actions. Senators questioned her previous defense of Edward Snowden, whom she had once praised for exposing illegal government activities. When pressed to label Snowden a traitor, she acknowledged that he broke the law but refrained from using the term “traitor.” 

    Regarding government surveillance, Gabbard had been a vocal critic of Section 702 of the Foreign Intelligence Surveillance Act (FISA), expressing concerns that citizens’ communications could be incidentally collected when targeting foreign nationals. 

    However, during the hearings, she indicated a shift in her stance, suggesting that with appropriate reforms, Section 702 could be a valuable tool for national security. This change led some Democrats to accuse her of political opportunism, while Republicans viewed it as a necessary evolution given her prospective role.

    Despite strong Democratic opposition, Gabbard’s nomination was confirmed by the Senate with a vote largely along party lines. 

    As Director of National Intelligence, Gabbard now oversees all 18 U.S. intelligence agencies, including the CIA, NSA, and FBI. Her appointment raises pressing questions about the future of U.S. intelligence policy. Will she uphold her past calls for transparency and civil liberties protections, or will she adopt a more traditional intelligence posture now that she’s at the helm?

    With rising global threats, cybersecurity challenges, and intense domestic political divisions, Tulsi Gabbard faces an uphill battle. The question is no longer whether she could get here. It’s whether she can succeed.