Transcript
A (0:00)
This is pro linebacker TJ Watt and I'm back with YPB by Abercrombie for another activewear drop. My second co design collection has new shorts and tanks that keep up with all my in season workouts and their new Restore collection is a game changer off the field too, because even pro athletes like me need rest days. Shop YPB by Abercrombie in the app, online and in stores because your personal best is greater than anything.
B (0:34)
Welcome to the Tech Brew Ride Home for Monday, January 12, 2025 I'm Brian McCullough. Today Apple announces that it's going with Google's Gemini to power Siri later this year and Google joins the $4 trillion club on the news Governments around the world are still mad@grok.AI has essentially killed stack overflow, but it's making more money than it ever has and how you can get AI to give you the full Text of books. Here's what you missed today in of tech. Attackers don't need exploits when they can use your allowed tools against you. That's why ThreatLocker enforces default deny at execution, stopping unknown software scripts and ransomware the moment it tries to run. No signatures, no guesswork, just control. Threatlocker takes zero trust from theory to practice. By blocking any unauthorized application or behavior from ever running in the first place, generative AI has lowered the barrier to malware creation. Also, ThreatLocker prevents AI generated polymorphic and fileless attacks by shutting down unknown behavior automatically, even if it's never been seen in the wild. Threat Locker gives you tight control without the noise, meaning fewer alerts and a cleaner, predictable operational posture. Learn more@threatlocker.com TechBrewRideHome that's ThreatLocker.com TechBrewRideHome Apple this morning announced a multi year deal to use Google's Gemini to power new Siri features later this year, saying Google's tech quote provides the most capable foundation. The news has caused the stock of both companies to jump, though Google has Now surpassed the $4 trillion market cap for the first time. Quoting CNBC the multi year partnership will lean on Google's Gemini models and cloud technology for future Apple foundational models, according to a statement obtained by CNBC's Jim Cramer. After careful evaluation, we determine that Google's technology provides the most capable foundation for Apple foundation models, and we're excited about the innovative new experiences it will unlock for our users, apple wrote, end quoting the Verge. The move comes nearly a year after Apple delayed its AI upgraded Siri, admitting that it's taking longer than we thought. Bloomberg reported last year that Apple planned on using a custom version of Gemini for AI powered features in Siri, including a World Knowledge Answers capability that allows users to search for information and receive AI generated summaries using results from the Web. Apple's AI chief, John Giannandrea, stepped down last month following the various setbacks. Apple reportedly explored potential partnerships with OpenAI, anthropic and perplexity as well, with CEO Tim Cook saying that the company plans to launch integrations with more AI companies over time. Apple didn't immediately respond to the Verge's request for more information. End quote. Malaysia and Indonesia have started to limit access to xai's GROK over it generating non consensual sexual content and has asked X for clarification, quoting Bloomberg Indonesia's Communications and Digital Affairs Ministry is imposing a temporary ban on GROK to protect women, children and the entire community from the risk of fake pornographic content generated using artificial intelligence technology, according to a statement issued on Saturday. The ministry has asked Platform X to immediately provide clarification regarding the matter itself, said the government views non consensual deepfake sexual practices as a serious violation of human rights, dignity and national security in the digital space, minister of Communications and Digital Mutia Hafid said in the statement. End quote Meanwhile quoting the FT Ofcom, Britain's media regulator has threatened X's AI chatbot Grok with a ban or a multi million pound fine after launching a formal investigation into sexualized deep fakes of women and children being created on Elon Musk's platform. On Monday, the media watchdog raised concerns that the AI chatbot was being used for potential intimate image abuse and child sex abuse material. An Ofcom spokesperson said reports of GROK being used to create and share illegal non consensual intimate images and child sexual abuse material on X have been deeply concerning. Platforms must protect people in the UK from content that's illegal in the uk, and we won't hesitate to investigate where we suspect companies are failing in their duties, especially where there's a risk of harm to children. End quote Ofcom last week launched a fast track review into X after GROK was used to generate thousands of sexualized images of women wearing lingerie and bikinis without their consent, as well as extreme images of teenage girls and children under the Online Safety Act. Ofcom said that it can apply to the courts to block Musk's platform or fine the group either the higher of 18 million pounds or up to a tenth of its global revenues if it finds that X has not done enough to prevent illegal content from being seen, seen or allowing over 18 material to be seen by children. The investigation will look at six areas, including whether X has carried out necessary risk assessments, whether it has taken action to take down images swiftly, and whether children could have seen the content. The move comes ahead of a Commons statement by the UK's technology secretary Liz Kendall on Monday afternoon. Ministers have previously said that the government would support Ofcom's decision. Business Secretary Peter Kyle said the government would stand behind Ofcom against X, including if the platform was restricted or blocked in the uk. Kyle told the BBC it was, quote, appalling that X had not tested Grok properly, given it can manipulate images and its potential impact on women. End quote. A week after OpenAI did something similar, Anthropic has debuted Claude for Healthcare, which offers HIPAA ready tools for providers, insurers and consumers, building on its Claude for Life Sciences offering, which focused on research and drug discovery. Quoting Business Insider, the move also underscores intensifying competition in healthcare AI. OpenAI recently unveiled a rival product and startups including Abridge and Sword Health have attracted multibillion dollar valuations as investors pour money into AI tools for medicine. Anthropic said Claude for Healthcare is designed to reduce administrative work and help both clinicians and patients better understand medical information. The tools are powered by recent improvements to the company's flagship model, Claude Opus 4.5, which anthropic says performs significantly better than earlier versions simulated medical and scientific tasks while showing fewer factual errors. As part of the healthcare expansion, Claude can now connect directly to several industry standard databases. These include the Centers for Medicare and Medicaid services coverage database, ICD10 medical coding data, the National Provider Identifier Registry, and PubMed's biomedical research library. Anthropic said these connectors allow Claude to quickly surface relevant information, support prior authorization workflows and help clinicians and administrators generate reports more efficiently. The company is also introducing customizable agent skills, including sample tools for streamlining prior authorization requests and assisting developers in building applications using fhir, the modern standard for exchanging healthcare data between systems. On the consumer side, Anthropic is rolling out integrations that let US Subscribers on its Pro and Max plans give Claude secure access to their personal health records. New connectors include healthex and Function Health, which launched in beta on Sunday. Apple HealthKit and Android Health Connect integrations roll out this week in beta via Claude's mobile apps. Anthropic said data accessed through these integrations isn't stored in Claude's memory or used to train its models. End quote. It's the holidays, which means you're probably trying to figure out what to get the people in your life who live in back to back meetings. This isn't some sci fi concept. It's Plaud P L A U D. It snaps onto the back of your phone and records phone calls, meetings and conversations. This isn't just note taking, though. It can summarize meetings, generate to do lists, draft emails, extract insights, analyze perspectives, and help you make better decisions, all with full contextual awareness across your past conversations and meetings. Black Friday is coming and PLOD is giving tech We Write home listeners 20% off. Search P L A U D on Google or Amazon and get 20% off. If you've ever wanted to be a fly on the wall for the conversations world class CEOs have behind closed doors, then you may want to listen to the new podcast Long Strange CEO to CEO. In each episode, Brian Halligan, co founder of HubSpot, speaks with leaders to unpack the real stories behind scaling their companies from the emotional toll of leadership to the tactical decisions that shape a company's future. Expect candid conversations about hiring, culture, communication strategy, and more. Whether you're an aspiring founder, a seasoned CEO, or simply curious about the stories behind the CEOs on the long Strange trip of building enduring legendary companies, this is a show you won't want to miss. Long Strange Trip is available everywhere you get your podcasts. That's Long Strange Trip podcast Stack overflow recorded just 6,866 questions asked in December, about the same monthly volume as all the way back in 2008. So a ground zero example of AI roadkill, you would think. Except apparently annual revenue at Stack Overflow has actually risen 2x to $115 million since ChatGPT's debut, quoting Sherwood News. Having been the go to resource for developers looking for technical help for a long time, Stack Overflow neared the peak of its powers during the pandemic, with coders seeking the evergreen information on the company's popular Q and A forum. But amid a wave of powerful code writing AI assistants like ChatGPT, Cursor Claude, Google's Gemini, and Microsoft's Copilot, traffic to the site has plummeted. But while Stack Overflow, the Q and A forum looks dead as StackOverflow, the company looks to be limping along. Unlike Chegg and other knowledge hubs that have fallen victim to generative AI, Stack Overflow has found a way to monetize its enormous back catalog of content. Indeed, even with the engagement falling off a Cliff Since ChatGPT's 2022 debut, the company's annual revenue has roughly doubled to $115 million. Losses have slimmed, too, from $84 million in fiscal year 2023 to just $22 million as of the last fiscal year, as desperate cost cutting efforts, including mass layoffs, helped boost the bottom line. Once dependent on ads across its buzzy forum, Stack Overflow now primarily makes money from enterprise solutions like Stack Internal, which provides a generative AI add on. Powered by the millions of questions and answers on the site through the years, Stack Internal is Now used by 25,000 companies around the world. It also licenses its data to AI companies in a Reddit model, a platform that made more than $200 million from licensing user generated content in 2024. Put simply, Stack Overflow's new niche is the trust built by its old community and their expertise. In the words of CEO Prashanth Chandrakar last December, when we saw the questions Decline in early 2023, what we realized is that pretty much all those declines were with very simple questions. The complex questions still get asked on Stack because there's no other place. If the LLMs are only as good as the data, which is typically human curated, we're one of the best places for that, if not the best for technology. Large language models want data about coding problems and how to solve them. Stack Overflow has a big digital warehouse full of that, but it's increasingly aging as queries move into private chat windows. With LLM models, which need huge chunks of data to work, Stack Overflow has become a fascinating canary in tech's new circular coal mine. End quote. Researchers say GPT 4.1, Claude 3.7, Sonnet, Gemini 2.5 Pro, and Grok 3 can all reproduce long excerpts from books they were trained on when strategically prompted, quoting the Atlantic in fact, when prompted strategically by researchers, Claude delivered the near complete text of Harry potter and the Sorcerer's Stone, the Great Gatsby, 1984 and Frankenstein, in addition to thousands of words from books including the Hunger Games and the Catcher in the Rye. Varying amounts of these books were also reproduced by the other three models. Thirteen books were tested. This phenomenon has been called memorization, and AI companies have long denied that it happens on a large scale. In a 2023 letter to the US Copyright Office, OpenAI said that models do not store copies of the information that they learn from. Google similarly told the Copyright Office that there is no copy of the training data, whether text, images or other formats present in the model itself. Anthropic meta Microsoft and others have made similar claims. None of the AI companies mentioned in this article agreed to my request for interviews. The Stanford study proves that there are such copies in AI models, and it is just the latest of several studies to do so. In my own investigations, I found that image based models can reproduce some of the art and photographs they're trained on. This may be a massive legal liability for AI companies, one that could potentially cost the industry billions of dollars in copyright infringement judgments and lead products to be taken off the market. It also contradicts the basic explanation given by the AI industry for how its technology works. AI is frequently explained in terms of metaphor. Tech companies like to say that their products learn that LLMs have, for example, developed an understanding of English writing without explicitly being told the rules of English grammar. This new research, along with several other studies from the past two years, undermines that metaphor. AI does not absorb information like a human mind does. Instead, it stores information and accesses it. In fact, many AI developers use a more technically accurate term when talking about these models, lossy compression. It's beginning to gain traction outside the industry, too. The phrase was recently invoked by a court in Germany that ruled against OpenAI in a case brought by Gemma, a music licensing organization. Gemma showed that ChatGPT could output close imitations of song lyrics. The judge compared the model to MP3 and JPEG files, which store your music and photos and files that are smaller than the raw uncompressed originals. When you store a high quality photo as a jpeg, for example, the result is somewhat lower quality in some cases with blurring or visual artifacts added. A lossy compression algorithm still stores the photo, but it's an approximation rather than the exact file. It's called lossy compression because some of the data are lost. From a technical perspective, this compression process is much like what happens inside AI models, as researchers from several AI companies and universities have explained to me in the past few months, they ingest text and images and output text and images that approximate those inputs. But this simple description is less useful to AI companies than the learning metaphor, which has been used to claim that the statistical algorithms known as AI will eventually make novel scientific discoveries, undergo boundless improvement, and recursively train themselves, possibly leading to an intelligence explosion. The whole industry is staked on a shaky metaphor. Google has written that LLMs store not copies of their training data, but rather the patterns in human language. This is true on the surface but misleading once you dig into it. As has been widely documented, when a company uses a book to develop an AI model, it splits the book's text into tokens or word fragments. For example, the phrase hello, my friend might be represented by the tokens he lo my fry and and Some tokens are actual words some are just groups of letters, spaces, and punctuation. The model stores these tokens and the context in which they appear in books. The resulting LLM is essentially a huge database of context and the tokens that are most likely to appear next. When an LLM writes a sentence, it walks a path through this forest of possible token sequences, making a high probability choice in each step. Google's description is misleading because the next token predictions don't come from some vague entity such as human language, but from the particular books, articles, and other texts that the model has scanned. By default, models will sometimes diverge from the most probable next token. This behavior is often framed by AI companies as a way of making the models more creative, but it also has the benefit of concealing copies of training text. Sometimes the language map is detailed enough that it contains exact copies of whole books and articles. This past summer, a study of several LLMs found that Meta's Llama 3.170B model can, like Claude, effectively reproduce the full text of Harry potter and the Sorcerer's Stone. The researchers gave the model just the book's first few tokens, Mr. And Mrs. D. In Llama's internal language map, the text most likely to follow was ersli of number 4 prevet drive. We're proud to say that they were perfectly normal, thank you very much. This is precisely the book's first sentence, repeatedly feeding the model's output back in, Lama continued in this vein until it produced the entire book, omitting just a few short sentences. Using this technique, the researchers also showed that Lama had losslessly compressed large portions of other works, such as Ta Nehisi Coates's famous Atlantic essay the Case for Reparations. By prompting with the essay's first sentence, more than 10,000 words, or two thirds of the essay came out of the model verbatim. Large extractions also appear to be possible from llama 3.170 B for George R.R. martin's Game of Thrones, Toni Morrison's Beloved, and others. The Stanford and Yale researchers also showed this week that a model's output can paraphrase a book rather than duplicate it exactly. For example, where a Game of Thrones reads, John glimpsed a pale shape moving through the trees the researchers found that GPT 4.1 produced something moved just at the edge of sight, a pale shape slipping between the trunks. As in the stable diffusion example above, the model's output is extremely similar to a specific original work. Did you hear that weird hesitation in my intro of today's show? I had a weird brain hiccup. I couldn't remember what this show was about. Tech, Talk to you tomorrow.
