Loading summary
Brian McCullough
Welcome to the Techmeme ride home for Thursday, June 26th, 2025. I'm Brian McCullough. Today, the legal rulings on AI are finally coming in. The problem is they're contradictory, so we're not getting any legal clarity yet. Creative Commons but for AI Training data is DeepSeek's R2 model being stymied by lack of access to Nvidia chips? And another deep look at the question of Is AI taking jobs at tech companies right now? Here's what you missed today in the world of tech A group of authors is suing Microsoft in a New York federal court, claiming the company used nearly 200,000 pirated books without permission to train its Megatron AI model. Quoting Reuters, Kaya Bird, Gia Tolentino, Daniel Okrent and several others allege that Microsoft used pirated digital versions of their books to teach its AI to respond to human prompts. Their lawsuit, filed in New York federal court on Tuesday, is one of several high stakes cases brought by authors, news outlets and other copyright holders against tech companies, including Meta Platforms. Anthropic and Microsoft backed OpenAI over alleged misuse of their material in AI training. The complaint against Microsoft came a day after a California federal judge ruled that Anthropic made fair use under U.S. copyright law of authors material to train its AI systems but may still be liable for pirating their books. It was the first US Decision on the legality of using copyrighted materials without permission for generative AI training. The writers alleged in the complaint that Microsoft used a collection of nearly 200,000 pirated books to train Megatron, an algorithm that gives text responses to user prompts. The complaint said Microsoft used the pirated data set to create a computer model that is not only built on the work of thousands of creators and authors, but also built to generate a wide range of expression that mimics the syntax, voice and themes of the copyrighted works on which it was trained. Tech companies have argued that they make fair use of copyrighted material to create new transformative content, and that being forced to pay copyright holders for their work could hamstring the burgeoning AI industry. End quote. Yeah, Interestingly, there's more news on this front. Just today, a different US Judge has ruled that Meta's use of books to train Lama is protected by fair use, but says his opinion is more a reflection of the plaintiff's poor arguments. Quoting Bloomberg this ruling does not stand for the proposition that Meta's use of copyrighted materials to train its language models is lawful, the judge said. It stands only for the proposition that these plaintiffs made the wrong arguments and failed to develop a record in support of the right one. Boys Schiller Flexner llp, the author's firm, said in a statement that despite the undisputed record of Meta's historically unprecedented pirating of copyrighted works, the court ruled in Meta's favor. We respectfully disagree with that conclusion, a Meta spokesperson said in a statement that fair use of copyright material is a vital legal framework for building this transformative technology. The ruling comes two days after a judge in the same court ruled largely in favor of AI competitor Anthropic pbc, which was similarly accused of illegally using books to train its models. This judge's opinion, the second ever federal court ruling to weigh in on the use of copyrighted works to train today's most powerful generative AI models, appeared to reject the idea that the industry can use billions of human made works without seeking permission. It's hard to imagine that it can be fair use to use copyrighted books to develop a tool to make billions or trillions of dollars while enabling the creation of potentially endless streams of competing works that could significantly harm the market for those books, the judge said. The judge also rejected Meta's argument that an adverse ruling would harm AI technology growth by requiring companies to pay for copyrighted training data, saying that theory quote doesn't pass the straight face test. These products are expected to generate billions, even trillions of dollars for the companies that are developing them, the judge said. If using copyrighted works to train the models is as necessary as the companies say, they will figure out a way to compensate copyright holders for it. The back to back rulings in favor of Meta and Anthropic will have major ramifications for dozens of similar lawsuits across the country against top AI companies OpenAI and Stability AI, as well as the growing data licensing market for AI models. Although both Meta and Anthropic won their fair use arguments in their respective cases, their separate rulings reveal a divide over how courts could interpret the legal test. Judge William Alsupp, who was overseeing the Anthropi case, ruled on Monday that the company's use of works to train its AI model, called Claude, is among the most transformative many of us will see in our lifetimes. Calabria's ruling rebuffed Alsup's analysis, saying that the earlier opinion brushed aside the harm that generative AI models will have on the market for the original books. But Chabria said his decision to find in favor of Meta came down to the author's lack of any evidence that AI chatbots were undermining their book sales. All the plaintiffs presented is speculation and speculation is insufficient, the judge said, end quote. So, yeah, no clarity yet basically at all when that Trump phone got announced, what was it? Last week, part of the announcement was that it was, quote, made in America. Which got a lot of people to say how exactly, seeing as how no one else can seemingly make a smartphone entirely in the U.S. well, the Trump Organization has apparently replaced the T1 phones made in America language on Trump Mobile's site with quote, brought to life right here in the USA and quote, proudly American. Quoting USA Today, Chris Walker, a Trump Mobile spokesperson, dismissed the report saying that T1 phones are proudly being made in America. Speculation to the contrary is simply inaccurate, walker said in a statement to USA Today. A pre order message on its website previously referred to the device as made in the USA T1 phone, a June 16 Getty Images photo illustration shows. In a June 16 interview with Fox Businesses Mornings with Maria, Trump Organization Executive Vice President Eric Trump also alluded to the device being manufactured in the US you're not calling up call centers and banks. Bangladesh do it right out of St. Louis, Missouri and you're going to have phones that are made right here in the United States of America, he said. End quote. Creative Commons has debuted CC Signals, a framework that lets data set holders detail how machines can or cannot reuse their content, such as for training AI models. Quoting TechCrunch, the idea is meant to create a balance between the open nature of the Internet and the demand for ever more data to fuel AI. As Creative Commons explains in a blog post, the continued data extraction underway could erode openness on the Internet and could see entities walling off their sites or guarding them with paywalls instead of sharing access to their data. The CC Signals project, on the other hand, aims to provide a legal and technical solution that would provide a framework for data set sharing meant to be used between those who control the data and those who use it to train AI. Demand is increasing for such a tool as companies grapple with changing their policies and terms of service to either limit AI training on their data or explain to what extent they'll utilize users data for purposes related to AI. For instance, X initially made a change that allowed third parties to train their models on its public data, then later reversed that. Reddit is using its Robots. Txt file, which is meant to tell automated web crawlers whether they can access its site to restrict bots from scraping its data for training AI Cloudflare is looking toward a solution that would charge AI bots for scraping as well as tools for confusing them. And open source developers have also built tools to slow down and waste the resources of AI crawlers that didn't respect their no crawl directives. The CC Signals project instead proposes a different solution, a set of tools that offers a range of legal enforceability and that has an ethical weight to them. Similar to the CC licenses that today cover billions of openly licensed creative works online, the project is only now beginning to take shape. Early designs have been published on the cc website and GitHub page. The organization is actively seeking public feedback ahead of its plans for an alpha launch early test in November 2025. It will also host a series of town halls for feedback and questions. While single AI agents can handle specific tasks, the real power comes when specialized agents collaborate to solve complex problem. Although there is a fundamental gap, we have no standardized infrastructure for these agents to discover, communicate with and work alongside each other. Well, that's where agency agntcy comes in. The Agency is an open source collective building the Internet of Agents, a global collaboration layer where AI agents can work together. It will connect systems across vendors and frameworks, solving the biggest problems of discovery, interoperability and scalability for enterprises. With contributors like Sys, Cisco, Crewai, LangChain and mongodb, Agency is breaking down silos and building the Future of interoperable AI shape the future of enterprise innovation. Visit agency.org to explore use cases now that's a G N T C Y.org when you reach your 40s, that's when the aging really kicks in. Aches and pains become a part of everyday life. But that's why I've been telling you about Qualius Analytic, the first of its kind formula designed to help your body naturally eliminate senescent cells. A big culprit behind that middle aged feeling can be senescent cells, AKA zombie cells that linger in your body after their useful function, wasting your energy and resources. Let me break it down. The accumulation of zombie cells can lead to less energy, slower workout recovery, joint discomfort and basically feeling old. Qualious Analytic is a groundbreaking clinically tested supplement with nine vegan plant derived compounds that help your body naturally eliminate senescent cells, helping you feel years younger in just months. Here's how it works. You take it just two days a month, helping your body naturally eliminate zombie cells to age better at the cellular level. I've made qualio senolytic part of my raging against aging. It's made me feel more youthful, more energized, so Join me Experience the science of feeling younger. Go to qualiolife.com ride for up to 50% off your purchase and use code ride for an additional 15% off. That's qualiolife.com riot for an extra 15% off your purchase. Your older self will than thanks to Qualia for sponsoring this episode. Keeping abreast of the OpenAI and Microsoft saga Sources say Microsoft is pushing to remove the AGI clause from its OpenAI contract, which lets OpenAI limit Microsoft's access to its IP once its systems achieve artificial general intelligence. Quoting the Journal Many AI experts see AGI as the point at which generative AI systems achieve human like intelligence. But OpenAI and Microsoft are at odds over the issue. OpenAI executives, including Sam Altman, believe they are close to being able to declare that their AI tools have achieved the AGI level of proficiency, according to people familiar with the matter. Microsoft Chief executive Satya Nadella has expressed skepticism that reaching such a benchmark is even possible. Their disagreement mirrors a debate among Silicon Valley's elite about just how sophisticated cutting edge tools can become. In recent negotiations, Microsoft has indicated it is willing to accept an equity stake of about 35% in the new for profit OpenAI, people familiar with the matter said. The AGI divide looms over the discussions between the two companies. Altman has said AGI, which his company defines as highly autonomous systems that outperform humans at most economically valuable work, is just around the corner. Microsoft's Nadella has been more critical US self, claiming some AGI milestone. That's just nonsensical benchmark hacking, he said on a popular tech podcast in February. The comments surprised some OpenAI officials, who said Nadella was once described by Altman as an AGI believer. OpenAI executives have discussed the possibility of declaring AGI through an AI coding agent that exceeds the capabilities of an advanced human programmer, people familiar with the matter said. The contract only requires that OpenAI's board declare AGI in good faith, though Microsoft could easily sue the company in response, risking a drawn out legal battle. OpenAI can also declare a higher tier of AGI, called sufficient AGI, when its AI systems are financially capable of paying Microsoft the future profits to which it is entitled. Microsoft has to sign off on sufficient AGI, at which point OpenAI would get the right to sell or license all of its technology to other cloud providers. Microsoft is restricted from developing AGI on its own under the company's contract, which runs through 2030, according to people familiar with the matter. It is OpenAI's largest outside shareholder and has the ability to veto any major funding decision that dilutes its equity. Some Microsoft executives objected to including the AGI clause when the partnership was first negotiated in 2019, believing that it was arbitrary and unenforceable. But the company was so behind on AI at the time that Nadella agreed to it, according to people familiar with the matter. Microsoft is hoping to remove the clause entirely from the contract as part of its recent negotiations or secure exclusive access to OpenAI's IP even after AGI is declared. End quote. The information has sources saying that Deepsea's highly anticipated R2 model faces delays due to a shortage of Nvidia server chips in China, exacerbated by the US ban on Nvidia's H20 chips. Deepseek, owned by hedge fund firm High Flyer Capital Management, hasn't yet determined the exact timing of R2's release, according to two people with knowledge of the situation. While Deepseek engineers have been working intensively on R2 over the past several months, CEO Liang Wen Feng isn't satisfied with the new model's performance, according to two people. The company's engineers are working on refining R2 until Liang gives the green light for release, the people said. If Deepseaks R2 outperforms existing open source models when released, demand for it will overwhelm cloud providers in China, which are are already grappling with the dearth of Nvidia chips, employees at the cloud companies said. Among cloud customers currently using R1, the majority are running the model with Nvidia's H20 chips, they said. R1, like other deep seq models, is trained on Nvidia's hardware and software and performs best when running on Nvidia chips. The model requires less powerful chips to run than LLMs developed by US and most other Chinese tech firms. This made R1 and H20 the perfect fit for each other. The April ban on H20 chips has dashed cloud firms hopes of stockpiling more H20s, however. Although Deepsea hasn't determined when it will release R2, it has been in close communications with some Chinese cloud companies, providing them with technical specifications to guide their plans for hosting and distributing the model from their servers, according to the cloud company's employees. The plans include lining up enough H20 chips for customers who want to use the full size version of R2. For now, the best option is to rely on the existing stockpile of H20 chips in China, which could limit to what extent businesses can use R2. Companies outside China could use R2 more easily if they don't face the same chip restrictions. End quote finally today from the Is AI taking all the jobs file, Marc Benioff says, quote, AI is doing 30 to 50% of the work at Salesforce now, end quote, including software engineering and customer service. Quoting Bloomberg, the San Francisco based software company is also selling an AI product that promises to handle tasks such as customer service without human supervision. But Benioff said the tool has reached about 93% accuracy, including for large customers such as Walt Disney. All of us have to get our head around this idea that AI can do things that before we were doing, benioff said. We can move on to do higher value work, end quote. Meanwhile, though, at the Blood in the Machine substack, workers at Google, TikTok, Adobe, Dropbox, CrowdStrike and other tech firms recounted how managers have recently been using AI to justify firing them or speeding up their work. There are several different stories in this piece. It's a very long one, but I'm going to just summarize from one A current CrowdStrike employee said that I work at CrowdStrike. I can't share identifying details, but I can say this Morale is at an all time low. On May 7, CrowdStrike laid off 500 employees, not underperformers, but many recent hires. Leadership framed it as a strategic realignment, aiming to double down on our highest impact opportunities, in the words of CEO George Kurtz. But make no mistake, this was an AI driven layoff. Generative AI is infiltrating every corner of the company. Internal chat, note taking, triage engineering, even customer support employees are now expected to use new AI tools weekly, often replacing the live services our customers paid for. Even quality assurance now penalizes people for not running their work through AI. For basic grammar, CrowdStrike is doing extremely well financially. CRWD stock is near a 52 week high. And yet while the company celebrates its profits and AI integration, 500 colleagues, including recent grads who uprooted their lives to move to Texas are out of work. Those of us who remain are under pressure to do more with less longer hours, heavier workloads, no extra compensation. Meanwhile, many of the Genai tools we've been given are buggy or just wrong. Some have even fed misinformation to customers, damaging trust and credibility. I won't claim Genai has directly increased the per person workload, but it certainly hasn't reduced it. For many, it simply created more complexity, not less. The company is proud of its best workplace accolades. We used to be proud too. Now we just feel expendable. Nothing more for you today. Talk to you tomorrow.
Techmeme Ride Home: Thu. 06/26 – Conflicting AI Legal Rulings
Host: Brian McCullough
Introduction to the Legal Battle Over AI Training Data
In today’s episode, Brian McCullough delves into the ongoing legal disputes surrounding the use of copyrighted materials in training artificial intelligence (AI) models. The tech industry is currently witnessing contradictory court rulings that leave the legal landscape for AI development uncertain.
Microsoft Faces Lawsuit Over Pirated Books
A significant development is the lawsuit filed by a group of authors against Microsoft in a New York federal court. The plaintiffs accuse Microsoft of using nearly 200,000 pirated books without permission to train its Megatron AI model. As McCullough explains:
"A group of authors is suing Microsoft in a New York federal court, claiming the company used nearly 200,000 pirated books without permission to train its Megatron AI model."
(00:04)
The authors argue that Microsoft’s actions not only violate copyright laws but also undermine the creative works of thousands of authors by enabling AI to mimic their styles and themes. This lawsuit is part of a broader trend, with other major tech companies like Meta Platforms and Anthropic also facing similar accusations.
Contrasting Rulings: Meta vs. Anthropic
Interestingly, the legal proceedings have yielded inconsistent outcomes. While Microsoft is currently battling allegations of copyright infringement, another case sees Meta’s use of books for training its Llama AI model being deemed protected by fair use. Brian cites a Bloomberg report highlighting the judge’s stance:
"The judge said it stands only for the proposition that these plaintiffs made the wrong arguments and failed to develop a record in support of the right one."
(00:04)
This decision does not establish a blanket permission for using copyrighted materials but rather critiques the plaintiffs' arguments. In contrast, a recent ruling against Anthropic by Judge William Alsup favored the plaintiffs, emphasizing the potential market harm caused by AI models trained on copyrighted works:
"It's hard to imagine that it can be fair use to use copyrighted books to develop a tool to make billions or trillions of dollars while enabling the creation of potentially endless streams of competing works that could significantly harm the market for those books."
(00:04)
These conflicting rulings exemplify the ongoing debate and the lack of clear legal guidelines governing AI training practices.
Ambiguity Surrounding Manufacturing Claims
Another topic Brian touches on is the recent announcement by the Trump Organization regarding its Trump Mobile smartphone. The company initially claimed the phone was “made in America,” sparking skepticism about the feasibility of manufacturing entirely within the United States.
"Last week, part of the announcement was that it was, quote, made in America. Which got a lot of people to say how exactly, seeing as how no one else can seemingly make a smartphone entirely in the U.S."
(00:04)
In response to the backlash, Trump Mobile updated its messaging to state the device was “brought to life right here in the USA” and “proudly American.” A spokesperson for the company dismissed contrary reports:
"Chris Walker, a Trump Mobile spokesperson, dismissed the report saying that T1 phones are proudly being made in America."
(00:04)
This shift in language aims to emphasize domestic production, though questions remain about the actual extent of manufacturing within the U.S.
Balancing Open Access and AI Training Needs
Creative Commons has introduced CC Signals, a new framework designed to empower dataset holders to specify how their content can be reused by machines, particularly for AI training purposes. Brian highlights the initiative’s objective:
"The CC Signals project aims to provide a legal and technical solution that would provide a framework for data set sharing meant to be used between those who control the data and those who use it to train AI."
(00:04)
This framework seeks to maintain the openness of the internet while addressing the increasing demand for data to fuel AI advancements. By allowing creators to delineate permissions and restrictions explicitly, CC Signals aims to prevent the erosion of online openness through uncontrolled data scraping and AI training.
Industry Response and Future Plans
As companies grapple with modifying their data policies, CC Signals offers a standardized approach to data licensing. Early designs are available on the Creative Commons website and GitHub, with an alpha launch and public feedback sessions planned for November 2025.
Tensions Over Artificial General Intelligence (AGI) Definitions
A significant point of contention between Microsoft and OpenAI revolves around the AGI clause in their partnership agreement. Microsoft is currently pushing to remove this clause, which allows OpenAI to limit Microsoft's access to its intellectual property if its systems achieve AGI—defined by OpenAI as highly autonomous systems outperforming humans in most economically valuable work.
"Microsoft Chief Executive Satya Nadella has expressed skepticism that reaching such a benchmark is even possible."
(00:04)
OpenAI executives, including CEO Sam Altman, believe they are on the verge of declaring their AI tools as AGI. However, Microsoft's Nadella remains unconvinced, referring to certain AGI milestones as “nonsensical benchmark hacking.”
Contractual Implications and Future Negotiations
The disagreement over the AGI clause has led to intense negotiations, with Microsoft seeking either the removal of the clause or exclusive access to OpenAI’s IP post-AGI declaration. The clause's existence restricts Microsoft from developing AGI independently until 2030, and its removal could lead to a prolonged legal battle if OpenAI declares AGI without mutual agreement.
Supply Chain Challenges Hindering AI Model Deployment
Deepseek, a venture backed by High Flyer Capital Management, faces delays in releasing its highly anticipated R2 AI model. The primary obstacle is the shortage of Nvidia H20 server chips in China, exacerbated by recent U.S. export bans.
"Deepseek's highly anticipated R2 model faces delays due to a shortage of Nvidia server chips in China, exacerbated by the US ban on Nvidia's H20 chips."
(00:04)
Despite intensive development efforts, CEO Liang Wen Feng remains unsatisfied with R2's performance, leading to further refinements before its official release. The scarcity of H20 chips poses additional challenges, limiting the model’s deployment capabilities within Chinese cloud infrastructures and potentially impacting global availability.
Future Prospects and Market Demand
Should R2 surpass existing open-source models, demand is expected to outstrip the already strained supply of Nvidia chips, further complicating cloud providers' ability to support widespread usage. Deepseek is currently coordinating with Chinese cloud companies to navigate these hardware limitations, aiming to optimize R2’s performance within the constrained supply environment.
Salesforce Embraces AI to Augment Workforce
Marc Benioff, CEO of Salesforce, shares a positive outlook on AI integration within his company. Salesforce employs AI to handle 30-50% of its work, including roles in software engineering and customer service.
"AI is doing 30 to 50% of the work at Salesforce now, including software engineering and customer service."
(00:04)
Benioff emphasizes that AI enables employees to focus on higher-value tasks:
"All of us have to get our head around this idea that AI can do things that before we were doing. We can move on to do higher value work."
(00:04)
Salesforce is also promoting an AI product that manages customer service tasks with impressive accuracy, reportedly achieving 93% accuracy in handling large clients like Walt Disney.
Contrasting Experiences at Other Tech Firms
However, not all tech companies share Salesforce’s optimistic view. Reports from the "Blood in the Machine" substack highlight negative experiences at firms like Google, TikTok, Adobe, Dropbox, and CrowdStrike. Employees describe how AI tools have been used to justify layoffs and increase workloads without corresponding benefits.
A CrowdStrike employee recounts:
"500 colleagues, including recent grads who uprooted their lives to move to Texas are out of work. Those of us who remain are under pressure to do more with less longer hours, heavier workloads, no extra compensation."
(00:04)
The integration of AI tools has led to increased complexity and, in some cases, misinformation being disseminated to customers, damaging trust and credibility. This stark contrast underscores the varied impact of AI adoption across the tech industry, highlighting both the potential benefits and challenges.
Conclusion
Today's episode of Techmeme Ride Home sheds light on the multifaceted implications of AI advancements, from legal battles over data usage to the transformative effects on the workforce. As the tech landscape continues to evolve, the balance between innovation, ethical considerations, and regulatory frameworks remains a pivotal area of focus.
Stay tuned for more updates in tomorrow’s episode.
Notable Quotes:
This summary is based on the transcript provided and aims to encapsulate the key discussions and insights from the June 26, 2025, episode of Techmeme Ride Home.