<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Kevin Bankston's CONVERGER]]></title><description><![CDATA[CONVERGER maps the content singularity where AI and the internet collapse all media into one: a connective node where emerging tech, policy, culture, futures thinking and storytelling intersect.]]></description><link>https://converger.kevinbankston.com</link><image><url>https://substackcdn.com/image/fetch/$s_!Kg3m!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54080bc3-3435-4e94-8a64-5ee48cb19baa_1024x1024.png</url><title>Kevin Bankston&apos;s CONVERGER</title><link>https://converger.kevinbankston.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 10 May 2026 22:28:04 GMT</lastBuildDate><atom:link href="https://converger.kevinbankston.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Kevin Bankston]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[kevinbankston@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[kevinbankston@substack.com]]></itunes:email><itunes:name><![CDATA[Kevin Bankston]]></itunes:name></itunes:owner><itunes:author><![CDATA[Kevin Bankston]]></itunes:author><googleplay:owner><![CDATA[kevinbankston@substack.com]]></googleplay:owner><googleplay:email><![CDATA[kevinbankston@substack.com]]></googleplay:email><googleplay:author><![CDATA[Kevin Bankston]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[CONVERGER #2 — YouTube’s Chilling Crackdown on AI Creators]]></title><description><![CDATA[Also: AI labs hypocritically condemn everyone else's internet scraping; viral internet creators dominate horror cinema; Noah Hawley is the franchise whisperer]]></description><link>https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown</link><guid isPermaLink="false">https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown</guid><dc:creator><![CDATA[Kevin Bankston]]></dc:creator><pubDate>Thu, 07 May 2026 14:02:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Kg3m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54080bc3-3435-4e94-8a64-5ee48cb19baa_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to <em>CONVERGER</em>, a biweekly newsletter mapping the content singularity where AI and the internet collapse all media into one&#8211;a connective node where emerging technology, policy, culture, futures thinking and storytelling intersect.</p><p><em>Converger </em>presents news and views from an AI, internet and media policy expert who is pro-innovation but anti-hype, allergic to both AI panic and AI boosterism, and passionate about supporting rather than supplanting human creativity with new technology.</p><p>Some issues may be heavier on media commentary, others on AI policy, others on personal passions like sci-fi&#8217;s influence on technology (both for good and bad) or the evolving medium and business of comic books in the digital age. You never know what threads might come together in convergence-space!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading CONVERGER! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I&#8217;m Kevin Bankston, your host. I&#8217;m an AI and internet law and policy expert, media nerd, and occasional fiction writer who works at a DC tech policy think tank and teaches AI and copyright law at a local law school. You can watch me develop newsletter content in real-time on <a href="http://linkedin.com/in/kevinbankston">LinkedIn</a> and the social network formerly known as <a href="https://x.com/KevinBankston">Twitter</a>, and less often on <a href="https://bsky.app/profile/bankston.bsky.social">Bluesky</a> and <a href="https://www.instagram.com/that_kevin_bankston/">Instagram</a>. You can also look for my more wonkish takes on AI governance at <a href="https://elicitation.substack.com/">Elicitation</a>, the new Substack from my AI policy day-job colleague Miranda Bogen of the Center for Democracy &amp; Technology&#8217;s AI Governance Lab. (Note that my Substack articles don&#8217;t necessarily reflect CDT&#8217;s positions.)</p><p>This week&#8217;s edition is unexpectedly YouTube-heavy: the video platform is a key player in the first two features and gets a big mention in the third. This week&#8217;s edition is also just <em>heavier</em>, generally: although I characterized the previous 8000-word issue as a &#8220;super-sized first edition,&#8221; this second one is closer to 10,000 words. Turns out I have a lot to say about what&#8217;s happening in the world of AI and content convergence!</p><p>So, I may be shifting from biweekly to weekly sooner rather than later to spread my content out. But the very next edition likely will not be out for three weeks, as I&#8217;ll be taking a vacation in the interim. Hopefully this super-<em>mega</em>-sized edition will tide you over until then.</p><p>So let&#8217;s go! And please share with your friends and colleagues if you enjoy!</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><h1>TABLE OF CONTENTS</h1><h2>FEATURES (&gt;500 words)</h2><ol><li><p><strong>&#8220;Deepfake&#8221; Detection Plus Demonetization Decimation Pose Double Threat to YouTube&#8217;s AI Creators (1527 words, 6 minute read)</strong></p></li><li><p><strong>The YouTube Generation Has Taken Over Horror Cinema (1546 words, 6 minute read)</strong></p></li><li><p><strong>Noah Hawley, the Multidimensional Franchise Expander (1545 words, 6 minute read)</strong></p></li><li><p><strong>Model Hypocrisy: Unconsented Data Scraping for Me But Not for Thee? (1435 words, 5.5 minute read)</strong></p></li><li><p><strong>A Hollywood AI Pipeline Built on Chinese Models? (537 words, 2 minute read)</strong></p></li><li><p><strong>How </strong><em><strong>Not</strong></em><strong> To Integrate AI Into Newsrooms: McClatchy&#8217;s AI-Driven Byline Blues and OpenAI&#8217;s Slop News Site (550 words, 2 minute read)</strong></p></li><li><p><strong>Hannah Einbinder Swirlie Watch: Who&#8217;s Getting Flushed for Using AI This Week? (595 words, 2.5 minute read)</strong></p></li></ol><h2>FRAGMENTS (&lt;500 words)</h2><ol><li><p><strong>Creator, Trademark Thyself</strong></p></li><li><p><strong>ChatGPT Images 2.0 Can Now Fake Your Doctor&#8217;s Note</strong></p></li><li><p><strong>Affleck Tops List of Most Powerful AI Players in Hollywood</strong></p></li><li><p><strong>New Creative AI Integrations, Integrating AI Creatively</strong></p></li><li><p><strong>Because Parasocial Relationships with AI Aren&#8217;t Weird Enough</strong></p></li><li><p><strong>Because Parasocial Relationships with AI Aren&#8217;t Weird Enough, Part Two</strong></p></li><li><p><strong>The </strong><em><strong>Other</strong></em><strong> Ryan Gosling Astronaut Movie</strong></p></li><li><p><strong>Warner Bros. Shareholders Approve Paramount-Skydance Merger Bid</strong></p></li><li><p><strong>Window Treatment: (Almost) All Studios To Let Movies Stay in Theaters Longer</strong></p></li><li><p><strong>Proving (or Pretending) You Didn&#8217;t Use AI for Your Writing</strong></p></li><li><p><strong>The Actors Guild Deal Has&#8230;Some Sort of AI Protections?</strong></p></li><li><p><strong>AI Film Wins at a Traditional Film Festival for First Time; Academy Nixes Awards for AI Writing and Performances</strong></p></li><li><p><strong>44% Slop and Rising: AI-Generated Music on Deezer</strong></p></li></ol><div><hr></div><h1>FEATURES</h1><h2>&#8220;Deepfake&#8221; Detection Plus Demonetization Decimation Pose Double Threat to YouTube&#8217;s AI Creators</h2><p>The same morning the first issue of <em>Converger</em> dropped, YouTube <a href="https://www.hollywoodreporter.com/business/digital/youtube-ai-deepfake-detection-tool-1236569593/">announced</a> that it is opening its AI-powered deepfake detection system to all of Hollywood. The tool will let any actor, athlete, musician, or creator who worries their likeness might be co-opted upload it to YouTube, get flagged when potential replicas appear, and request takedowns. The system has been quietly available to top creators and a select group of politicians and public officials for months; now it&#8217;s going wide. The talent agents quoted in <em>The Hollywood Reporter</em>&#8216;s exclusive coverage were practically giddy. Forgive me for not also applauding.</p><p>As YouTube&#8217;s Chief Business Officer admits in the story, this is basically like Content ID&#8211;YouTube&#8217;s tool for detecting copyrighted content&#8211;but for famous people&#8217;s faces. And as we all know, Content ID has <em><a href="https://www.eff.org/wp/unfiltered-how-youtubes-content-id-discourages-fair-use-and-dictates-what-we-see-online">never</a></em> interfered with legitimate speech.</p><p>I very much want to know more details about this new initiative, but the story is light on specifics. How will the tool distinguish between AI-generated and real-life images of public figures (if at all)? How will YouTube apply carveouts for news, criticism, parody, and satire, types of speech for which public figures are particularly legitimate targets? Based on what right or policy would takedown requests related to abusive use of replicated faces even be based on? What counts as &#8220;abuse&#8221; or a &#8220;replica,&#8221; anyway? Only YouTube knows.</p><p>Frankly, I don&#8217;t have high hopes for this not causing a lot of collateral damage to legitimate expression on YouTube&#8217;s platform based on what I&#8217;ve seen all over Twitter the past few weeks: endless complaints from creators using AI in their otherwise human-created YouTube videos (and many who don&#8217;t), who are being demonetized by YouTube&#8217;s automated processes to detect &#8220;inauthentic content&#8221; that violates its policies. If the apparently unjustified breadth of this crackdown is any indication, I&#8217;d expect the special protections for celebrity faces to be applied in an equally imprecise manner.<br><br>I&#8217;ve yet to see anything from the mainstream tech or media press on this demonetization issue, but it&#8217;s been absolutely dominating my Twitter feed. A few examples:</p><blockquote><p>Wow. So YouTube claims my content isn&#8217;t authentic despite the overwhelming majority of my videos being videos of myself on camera and my own music recorded live in a studio. @TeamYouTubeI&#8217;ve been a content creator with you for over a decade. What&#8217;s going on right now?&#8212;<a href="https://x.com/ThePholosopherX/status/2048633764796928394">@ThePholosopherX<br><br></a>I run 2 YouTube channels (714K and 350K subs) with more than 400 million views on long-form videos. I create comedy podcasts and gaming content, and I do everything myself: filming, editing, and appearing on camera. Overnight, both channels were demonetized. I made 2 appeal videos&#8212;zero views. Every appeal was rejected with the exact same copy-paste automated message. No human support from YouTube. After 12 years and building two big communities, this treatment is unacceptable. &#8211;@<a href="https://x.com/DeliresdeMax/status/2050531802620567782">DeliresdeMax</a></p><p>I&#8217;m a solo creator who runs the channel <em>Sacred Stuff</em>. It&#8217;s a cinematic storytelling channel where I post very low-volume, heavily detailed animated stories. I have over 80,000 subscribers and have only 4 videos. I was marked for inauthentic content. The policy specifies content that is mass-produced or repetitive. None of my stories are repetitive at all, and I only post once a MONTH, so they&#8217;re not mass-produced. Each video takes about 500 hours of human work. I walk through my entire process in this appeal video below[.] Can a human please review this?&#8212;<a href="https://x.com/sacredstuf/status/2050963229828853961">@sacredstuf</a> [UPDATE: it looks like this Twitter account has just been closed, and the creator has started a new account, <a href="https://x.com/sacredstuffYT">@sacredstuffYT</a>]</p><p>Seems there is a mass YouTube demonetization with channels being slapped as &#8217;Inauthentic Content&#8217;. Affecting both AI &amp; non AI creators. YT is not handling this AI wave well. I understand demonetizing those fully AI generated content farms that upload multiple times a day. But why take down animators, AI assisted storytellers &amp; even some completely non AI channels? Is this another youtube apocalypse?&#8212;<a href="https://x.com/AzeAlter/status/2040917741369168302">@AzeAlter</a></p></blockquote><p>Self-proclaimed &#8220;viral AI storyteller&#8221; <a href="https://x.com/BLVCKLIGHTai/status/2047335095095398667">BLVCKL!GHT</a> catalogued more affected creators:</p><blockquote><p>This week, several of us were demonetized by @TeamYouTube for &#8220;inauthentic content.&#8221; None of us received an explanation.</p><p>Myself (<a href="https://youtube.com/@blvcklightai?si=_ONM9ogeiKqBNbwi">https://youtube.com/@blvcklightai?si=_ONM9ogeiKqBNbwi</a>) is a two-time Escape Awards-winning AI filmmaker whose work has been exhibited internationally and across multiple streaming platforms.</p><p>Aze Alter (<a href="https://youtube.com/@azealter?si=JIvdVFeVvI1G1oQO">https://youtube.com/@azealter?si=JIvdVFeVvI1G1oQO</a>) is a writer and director behind some of the most recognized sci-fi worldbuilding series in the AI filmmaking space, with 230K+ YouTube subscribers.</p><p>Mart Zien (<a href="https://youtube.com/@azealter?si=JIvdVFeVvI1G1oQO">https://youtube.com/@azealter?si=JIvdVFeVvI1G1oQO</a>) is a Cannes-credited film producer whose AI work has swept major festivals, been covered in <em>Forbes</em>, and been cited on the <em>Masters of Scale</em> podcast. [Note the correct name is <em>Matt </em>Zien, and the correct URL is <a href="https://www.youtube.com/@kngmkrlabs">https://www.youtube.com/@kngmkrlabs</a>]</p><p>Javi Lopez (<a href="https://youtube.com/@javilopen?si=FU6UouE386mQAOOQ">https://youtube.com/@javilopen?si=FU6UouE386mQAOOQ</a>) is the founder of Magnific AI, a tool used in the VFX of a major Robert Zemeckis film starring Tom Hanks.</p><p>One search tells you exactly who we are and what we do.</p><p>So when @TeamYouTube and @YouTubeCreators label our work &#8220;inauthentic&#8221; with no definition, no breakdown, and no appeal path worth using, that&#8217;s not moderation. That&#8217;s a platform that has stopped being able to tell the difference between spam and craft.</p></blockquote><p>Notably, although this demonetization occurs because these videos are nominally unsuitable for advertising against, that <a href="https://x.com/DaveCullenShow/status/2049822153604800861">won&#8217;t actually stop</a> YouTube from continuing to stick advertisements on them while not giving creators a cut.</p><p>It&#8217;s also ironic to have Google promoting its state-of-the-art Veo video generation model with one hand, and then punishing AI creators on its platform with the other, an inconsistency captured well by <a href="https://x.com/brianjamesgage/status/2048470009773707708">one affected Twitter user</a> in graphic form:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!f7ij!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!f7ij!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 424w, https://substackcdn.com/image/fetch/$s_!f7ij!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 848w, https://substackcdn.com/image/fetch/$s_!f7ij!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 1272w, https://substackcdn.com/image/fetch/$s_!f7ij!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!f7ij!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png" width="1190" height="646" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:646,&quot;width&quot;:1190,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!f7ij!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 424w, https://substackcdn.com/image/fetch/$s_!f7ij!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 848w, https://substackcdn.com/image/fetch/$s_!f7ij!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 1272w, https://substackcdn.com/image/fetch/$s_!f7ij!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F408d3cd9-ba76-42d8-9efb-09d0f5d215c5_1190x646.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Meanwhile, YouTube&#8217;s primary message&#8211;rather than actually giving clearer guidance on what will get you demonetized, or more specific explanations to the creators about why they were demonetized&#8211;has essentially been <a href="https://x.com/TeamYouTube/status/2049798544974843963">&#8220;if you disagree, change your content or file an appeal.&#8221;</a> The same message was repeated in a more reassuring tone by the <a href="https://x.com/gstrompolos/status/2047728025132032370">co-founder of AI studio Promise</a>, who previously helped build YouTube&#8217;s monetization program:</p><blockquote><p>We checked in with YouTube about the demonetization issues impacting the Gen AI creator community. They are continuing efforts to reduce repetitive, spammy content so more original, story-driven work can rise. The system isn&#8217;t perfect, so if you were flagged incorrectly, appeal. That feedback helps improve things over time. There&#8217;s a wave of bold, original work coming from AI creators right now, and we will continue to advocate for you.</p></blockquote><p>And it does seem that the appeal process is working to some extent; a number of the louder voices that complained on Twitter, including <a href="https://x.com/BLVCKLIGHTai/status/2047735667988599139">BLVCKL!GHT</a>, have now been reinstated. That&#8217;s both good and bad: good that many mistakes have been corrected; bad that so very many were made in the first place. Clearly, YouTube&#8217;s automated censors are not well-calibrated and these digital dragnets are catching a lot more than intended.</p><p>At least some of those creator-complainants were savvy enough to <a href="https://x.com/javilopen/status/2048319298335097234">specifically call out</a> the European Union&#8217;s Digital Services Act in their appeals. That law requires a specific explanation of demonetization or deplatforming decisions by user-generated content platforms, in addition to requiring an appeals process. I&#8217;ve now seen a lot of people <a href="https://x.com/DaveCullenShow/status/2048681081990181020">reposting</a> their notices from YouTube and they certainly don&#8217;t feel very specific&#8211;just restatements of the very vague, broad, subjective provision in the terms of service against &#8220;inauthentic content,&#8221; which originated from an <a href="https://support.google.com/youtube/answer/1311392?hl=en#zippy=%2Cfollow-the-youtube-community-guidelines%2Cfollow-our-program-policies">expansion</a> of the &#8220;repetitious content&#8221; policy in July 2025:</p><blockquote><p>Inauthentic content refers to mass-produced or repetitive content. This includes content that looks like it&#8217;s made with a template with little to no variation across videos, or content that&#8217;s easily replicable at scale&#8230;. Examples of what&#8217;s not allowed to monetize (this list is not exhaustive):</p><ul><li><p>Content that exclusively features readings of other materials you did not originally create, like text from websites or news feeds</p></li><li><p>Songs modified to change the pitch or speed, but are otherwise identical to the original song</p></li><li><p>Similar repetitive content with low educational value, commentary, narratives, or minimal variation across videos</p></li><li><p>Mass-produced content using a similar template across multiple videos</p></li><li><p>Image slideshows or scrolling text with minimal or no narrative, commentary, or educational value</p></li></ul></blockquote><p>To the extent I understand what the above includes, it seems that many of the folks complaining don&#8217;t fit the bill&#8211;and apparently YouTube (eventually) agrees, considering many of the reinstatements.</p><p>YouTube certainly has a right to make choices about what&#8217;s on its platform, and there is indeed a lot of AI (and non-AI) engagement-farming garbage that it can and should work to get rid of. But right now it&#8217;s clearly getting the balance wrong&#8211;either in terms of how it has written its &#8220;inauthentic content&#8221; policy, how it is interpreting it, how it is automatically applying it, how it is explaining those decisions, how it is handling appeals to those decisions, or all of the above.</p><p>YouTube needs to take a beat to pause this decimating wave of demonetization, take another look at its DSA obligations and the <a href="https://santaclaraprinciples.org/">Santa Clara Principles on Transparency and Accountability in Content Moderation</a> that helped inspire them, and go back to the drawing board. It should be equally cautious and much more forthcoming about its deepfake detection initiative, as well. Otherwise YouTube risks stifling the very same AI-driven creativity they are promising to enable with Google&#8217;s own AI models, and pushing the next generation of tech-enabled talent to use another platform.</p><h2>The YouTube Generation Has Taken Over Horror Cinema</h2><p>The poorly-targeted wave of demonetization currently hitting YouTube video creators takes on additional resonance when you look at how important that platform has been to cultivating the film talent of the future. In our <a href="https://kevinbankston.substack.com/p/ed2d73d3-f032-4b23-b34c-d354b6c11a33">last issue</a>, I highlighted the upcoming feature film <em>Backrooms</em> as the latest product of the YouTube-creator-to-Hollywood-director pipeline. Several items came across my TL since then to highlight that not only is that pipeline flowing, it&#8217;s near bursting with fresh talent. But for how long?</p><p>The first item was this <a href="https://x.com/sillierdeadite/status/2047064325907476836">meme-chart</a>, highlighting a huge roster of feature horror directors with YouTube roots:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LVqg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LVqg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 424w, https://substackcdn.com/image/fetch/$s_!LVqg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 848w, https://substackcdn.com/image/fetch/$s_!LVqg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 1272w, https://substackcdn.com/image/fetch/$s_!LVqg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LVqg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png" width="954" height="1352" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1352,&quot;width&quot;:954,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LVqg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 424w, https://substackcdn.com/image/fetch/$s_!LVqg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 848w, https://substackcdn.com/image/fetch/$s_!LVqg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 1272w, https://substackcdn.com/image/fetch/$s_!LVqg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F92fb7325-c208-454e-a8c3-55e5408dfc56_954x1352.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The marquee example is <strong>Kane Parsons</strong> (AKA Kane Pixels), director of <em>Backrooms</em>, whose story is illustrative of this internet trend not just because he&#8217;s a YouTube creator but one who made his bones expanding on creepypasta internet lore and relied on design software to render it.</p><p>The concept of the <em>Backrooms </em>grew out of an unattributed photo and text snippet on 4chan and blossomed into an entire subculture of creators telling stories based in a horrific and endless liminal space of empty, sickly-yellow office hallways, an other-dimensional labyrinth pulsing with mundane dread.</p><p>Parsons became the preeminent chronicler of the <em>Backrooms</em>. His <a href="https://www.youtube.com/watch?v=H4dGpz6cnHo">debut video</a> in January 2022 was made by the sixteen year-old Parsons in a month using open source 3D modeling software Blender and Adobe After Effects, and it and its <a href="https://www.youtube.com/playlist?list=PLVAh-MgDVqvDUEq6qDXqORBioE4Yhol_z">over twenty sequels</a> have collectively been viewed nearly two hundred million times. And now, at twenty years old, he&#8217;s the youngest director of any feature film released by A24, Hollywood&#8217;s preeminent purveyor of elevated horror films. (And, on a tech side-note, he <a href="https://www.hollywoodreporter.com/movies/movie-news/backrooms-kane-parsons-youtube-a24-horror-movie-ccxp-1236577326/">used Blender extensively</a> for the feature film too.)</p><p>Next up in the chart is Curry Barker, whose debut feature&#8211;love spell gone wrong horror flick <em>Obsession</em>&#8211;is coming to theaters on May 15. Until last week he was the second-biggest name-to-watch on this list, but today he is arguably the most. Barker came up through <em><a href="https://www.youtube.com/@thats_a_bad_idea">That&#8217;s a Bad Idea</a></em>, the YouTube comedy sketch channel he runs with Cooper Tomlinson (favorably compared to <em>I Think You Should Leave</em>), then dropped the<a href="https://www.youtube.com/watch?v=pbzGQ1lszv4"> $800 viral horror short </a><em><a href="https://www.youtube.com/watch?v=pbzGQ1lszv4">Milk &amp; Serial</a></em> on YouTube in 2024. A few days ago he wrapped principal photography on his next horror pic for Focus,<a href="https://deadline.com/2026/04/aaron-paul-curry-barker-anything-but-ghosts-focus-features-1236764637/"> </a><em><a href="https://deadline.com/2026/04/aaron-paul-curry-barker-anything-but-ghosts-focus-features-1236764637/">Anything But Ghosts</a></em>, produced by Blumhouse-Atomic Monster and Spooky Pictures, <em>and </em>he just signed on to direct <a href="https://variety.com/2026/film/news/texas-chainsaw-massacre-curry-barker-director-obsession-1236727385/">A24&#8217;s </a><em><a href="https://variety.com/2026/film/news/texas-chainsaw-massacre-curry-barker-director-obsession-1236727385/">Texas Chainsaw Massacre</a></em><a href="https://variety.com/2026/film/news/texas-chainsaw-massacre-curry-barker-director-obsession-1236727385/"> reimagining</a> also produced by Spooky Pictures. Three features in the pipeline and the keys to a major IP, all before his theatrical debut.</p><p>The remaining examples are equally impressive:</p><p><strong>Danny &amp; Michael Philippou: </strong>Creators of <em><a href="https://www.youtube.com/channel/UCz_cDc_2arKIb6SlJoqFT0w">RackaRacka</a></em> (6.9M subs), an Australian comedy channel built on hyper-violent practical-effects sketches, the Philippou brothers directed 2023&#8217;s <em>Talk to Me</em> and their 2025 follow-up <em>Bring Her Back</em>, both for A24.</p><p><strong>Chris Stuckmann: </strong>Stuckmann spent over fifteen years as one of YouTube&#8217;s most respected <a href="https://www.youtube.com/@ChrisStuckmann">film critics</a> (2M subs), then jumped from reviewing horror to making it. The Kickstarter raised $1.4M. Mike Flanagan came on as EP. Neon distributed.</p><p><strong>Dan Trachtenberg: </strong>The granddaddy of this list. His 2011 live-action <em><a href="https://www.youtube.com/watch?v=4drucg1A6Xk">Portal: No Escape</a></em> fan short, made on a shoestring and posted to YouTube, got him <em>10 Cloverfield Lane</em>, and that got him <em>Prey</em> and now <em>Predator: Badlands</em>. He&#8217;s proof that this pattern long predates the current wave.</p><p><strong>Michael Shanks: </strong><em><a href="https://www.youtube.com/timtimfed">timtimfed</a></em>, Shanks&#8217; Australian YouTube comedy channel of surreal genre shorts, ran for years before his debut body-horror feature <em>Together</em>, produced by its stars Alison Brie and Dave Franco, broke at Sundance and grossed $32M worldwide last year.</p><p><strong>Mark &#8220;Markiplier&#8221; Fischbach: </strong>38M-sub gaming YouTuber Fischbach first played <em>Iron Lung</em>, the David Szymanski indie sci-fi horror game, on his stream in May 2022. He then optioned and self-financed the film adaptation, clearing over $50M worldwide earlier this year on a $3M budget and without traditional studio backing.</p><p>An illustrious roster, but what&#8217;s most notable about this meme is that based on news from the past few weeks it is already out of date, since there are at least four more creators of viral internet media with recent news about their burgeoning directorial careers:</p><p><strong>Sam Evenson: </strong>Evenson is a VFX artist (<em>Dune: Part Two</em>, <em>The Last of Us</em>) who runs the YouTube channel <em><a href="https://www.youtube.com/@GrimoireHorror">Grimoire Horror</a></em>. His viral 12-minute YouTube short <em><a href="https://www.youtube.com/watch?v=mUXpTAdu_8U">Mora</a></em>, about a haunted image generation AI model, <a href="https://variety.com/2026/film/news/neon-adapting-sam-evenson-viral-ai-horror-short-mora-feature-film-1236727252/">sold to Neon</a> two weeks ago, with Evenson writing and directing the feature. Like Barker, his production is backed by Spooky Pictures&#8217; Steven Schneider and Roy Lee. Notably, Lee also was one of the producers who launched the feature career of <strong>Zach Cregger</strong> (<em>Weapons</em>, the upcoming <em>Resident Evil</em> reboot) with his breakout debut <em>Barbarian.</em> (Like horror director-producer Jordan Peele and many others on this list, Cregger started in comedy before transitioning to horror, as a writer-performer in the sketch group <em>The Whitest Kids U&#8217;Know</em>.)</p><p><strong>Ian Tuason: </strong>Tuason has the weakest YouTube connection on this list, but <a href="https://x.com/ErikDavis/status/2046986687184621772">Erik Davis</a> of <em>Rotten Tomatoes</em> and <em>Fandango</em> grouped him with the others in a recent tweet and he&#8217;s not wrong: &#8220;It&#8217;s interesting that A24 is not-so-quietly recruiting filmmakers like Kane Parsons, Curry Barker, the Philippou brothers, Ian Tuason&#8212;all guys who created viral content prior to their big-screen debuts.&#8221; In this case, Tuason made indie shorts through the 2010s; his 2025 horror feature <em>Undertone</em> started life as a radio play, was shot for $500K in his childhood home, won the Fantasia audience award, sold to A24 in a mid-seven-figure bidding war, and pulled a $9.3M domestic opening weekend. A month ago he confirmed in<a href="https://www.phantasmag.com/articles-2/undertone-sequel-exclusive-ian-tuason-interview"> an interview with </a><em><a href="https://www.phantasmag.com/articles-2/undertone-sequel-exclusive-ian-tuason-interview">Phantasmag</a></em> that A24 will be producing an <em>Undertone</em> prequel, and a third entry is being discussed. He is also lined up to direct <em><a href="https://bloody-disgusting.com/interviews/3939766/ian-tuason-teases-his-paranormal-activity-reboot-exclusive/">Paranormal Activity 8</a></em> for Blumhouse.</p><p><strong>Casper Kelly: </strong>Kelly&#8217;s the delayed-action case. His <em><a href="https://www.youtube.com/watch?v=QrGrOK8oZG8">Too Many Cooks</a></em>&#8212;a horrifyingly surreal 11-minute Adult Swim sitcom parody from 2014&#8212;became one of the defining viral shorts of the 2010s after escaping its late-night cable slot onto YouTube and Vimeo. Twelve years later, that viral artifact is getting him his feature directorial debut. <em>Buddy</em>, a Sundance Midnight standout about a tyrannical children&#8217;s TV unicorn, was <a href="https://x.com/KevinBankston/status/2047479877519946005">acquired two weeks ago</a> by Roadside Attractions and Saban Films for a Labor Day wide release. The pipeline doesn&#8217;t always run on the four-year arc Parsons modeled; sometimes a viral artifact sits in the cultural memory for a decade before the producers that are game to back that auteur show up. (Buddy is produced by BoulderLight Pictures&#8212;the indie horror shingle that launched Zach Cregger&#8217;s career with <em>Barbarian</em>&#8212;alongside Low Spark Films.)</p><p><strong>Dylan Clark: </strong>As<strong> </strong><a href="https://deadline.com/2026/04/the-blair-witch-project-reboot-details-revealed-fall-shoot-1236876959/">announced just last week</a>, this YouTube horror short director (his <em><a href="https://www.youtube.com/watch?v=BI9fKfX5V68">Portrait of God</a></em> has nearly 9M views) has been tapped for the <em>Blair Witch</em> reboot from Lionsgate and Blumhouse at a $10M budget. Clark also has <em>Portrait of God</em> in development as a feature at Universal with Sam Raimi and Jordan Peele producing, and a third feature <em>Story Time</em> set up at LD Entertainment. Three projects, one debut feature, and the two most prestigious horror producers alive in his corner. It&#8217;s also worth flagging that Steven Schneider&#8211;the Spooky Pictures producer behind Curry Barker&#8217;s <em>Texas Chainsaw</em> and Sam Evenson&#8217;s <em>Mora</em>&#8211;is also EP on <em>Blair Witch</em>.</p><p>Having one of these YouTube kids directing a Blair Witch flick seems like a full-circle moment, since that film&#8217;s low-budget found-footage aesthetic and its viral internet lore-marketing were such clear influences on projects like <em>Backrooms</em>.</p><p>All of the above demonstrates a few takeaways about this internet-to-theaters horror trend (a trend that just before press time on May 6th was also <a href="https://theankler.com/backrooms-kane-parsons-obsession-curry-barker-young-horror/">reported on by Matthew Frank at the Ankler</a>, with a focus on Parson, Barker, and Clark):</p><ul><li><p>First, this pipeline isn&#8217;t a new phenomenon but a mature one, and between A24, the Blumhouse folks, and Spooky Pictures, there is now a substantial contingent of star horror producers who clearly have made the YouTube-to-feature handoff a primary deal-making lane. That trend is likely to continue and expand. And although A24 is still the studio leading in YouTube-to-silver-screen conversions, all the other studios are starting to get in the game.</p></li><li><p>Both comedy and horror are the pathway, but horror is always the feature destination. Why? Two structural forces are doing the work. Genre:<strong> </strong>horror is the last mass-market feature category Hollywood still makes both cheap and often&#8212;<em>Iron Lung</em> on $3M, <em>Obsession</em> shot for under $1M, <em>Shelby Oaks</em> on $2.8M. There is no equivalent low-budget pipeline for drama or comedy. Craft adjacency: horror and comedy are both nervous-system genres, utilizing the same skills of tension build and release that can lead either to a belly laugh or a fearful scream or both. Someone who has spent a hundred videos learning how to land a sketch beat or a jump scare on a thirteen-year-old&#8217;s algorithm-addled attention has been training the exact muscle the genre rewards.</p></li><li><p>It&#8217;s all dudes. Almost all white dudes, except for Filipino-Canadian Ian Tuason and half-Korean Markiplier. I refuse to believe there aren&#8217;t plenty of equally talented women and people of color jamming out YouTube content that demonstrates their capability to take on a small-budget horror pic. Producers and studios need to go find them.</p></li></ul><p>And the final takeaway: YouTube is cracking down on AI-driven creators at its peril, risking the next generation of Kane Parsons-style innovators who use the most available technology to quickly ship their most viral ideas to a hungry audience. Yes, there are slop-meisters that need to be demonetized, but there are also budding auteurs who may never flower, or will move to flourish on another platform, if YouTube can&#8217;t calibrate its censor-bots to distinguish the two.</p><h2>Noah Hawley, the Multidimensional Franchise Expander</h2><p>The default mode of Peak IP is extraction. Studios hold a vault of recognizable names, and the job of each new season, sequel, or spin-off is to get more ore out of the same shaft: same story beats, same characters, same visual language, diminishing returns.</p><p>We have a decade of receipts now on what that looks like. Cinema and TV are dominated by nostalgia reboots, legacy sequels, cinematic universes that can no longer remember why they exist. And we&#8217;re about to find out what it looks like when generative AI makes the cost of extraction approach zero. Slop strip-mining of old IP, be it authorized or unauthorized, is the endpoint of the extraction model.</p><p>Which is why it&#8217;s worth naming, precisely, what Noah Hawley&#8211;the TV showrunner for FX&#8217;s <em>Fargo</em>, <em>Legion</em>, and <em>Alien: Earth</em>&#8211;is doing instead.</p><p>Two things recently put Hawley back in front of me. The first was <a href="https://www.theatlantic.com/magazine/2026/05/billionaire-consequence-free-reality/686588/">his essay</a> in <em>The Atlantic</em> on the consequence-free reality of the contemporary billionaire, based on his visit to Jeff Bezos&#8217; &#8220;Campfire&#8221; retreat; it was a big part of The Internet Discourse a couple weeks ago, and also offered an interesting lens with which to view the billionaire villain in his<em> Alien </em>series.</p><p>More interesting to me, though, was <a href="https://variety.com/2026/tv/global/fargo-alien-earth-noah-hawley-youtube-canneseries-1236730351/">his Canneseries interview</a>, also at the end of April, in which Hawley repeated a label someone else hung on him recently: &#8220;franchise whisperer.&#8221;</p><p>The label is fine. But it&#8217;s not precise enough. What I see Hawley doing across <em>Fargo</em>, <em>Legion</em>, and <em>Alien: Earth</em> isn&#8217;t just coaxing new life out of old IP. It&#8217;s a structural method, and once you see it, you can&#8217;t unsee it.</p><p>I would label Hawley as a <strong>three-dimensional franchise expander</strong>: someone who extends other people&#8217;s properties along distinct axes, and who picks a different axis each time depending on what the source material is actually asking for.</p><p>One might ask, why focus on expanding other people&#8217;s source material at all? Hawley is a uniquely gifted writer-creator who certainly has the capacity to craft his own original stories; he&#8217;s already done so repeatedly as a novelist.</p><p>The answer is in the same Canneseries conversation, where Hawley candidly described the competitive landscape for TV. His biggest competition, he said, is YouTube: a platform that spends nothing to produce content while he&#8217;s spending $250 million per season of <em>Alien: Earth</em>. When the cost of content collapses to zero, recognizable brands become the lifelines that pull eyeballs back to expensive storytelling. Which is why how you build on those brands&#8211;extractive vs. additive, mining vs. growing&#8211;becomes the central craft question of the infinite-content era.</p><p>Hawley&#8217;s craft is not extracting, it&#8217;s expanding. Or, to cite an old bit of advice from comic book writers, the original shared-fictional-universe creators: when you take a toy out of the franchise toybox for your run on a comic series, you need to make sure you put it back right when you&#8217;re done&#8211;and ideally add a few new toys to the box for the next creator to play with.</p><p>Noah Hawley always creates new toys. Watch him work along three axes:</p><h4>FORWARD: <em>Fargo</em></h4><p>The easy thing to do with the Coens&#8217; 1996 film would have been to retell it at a more leisurely and detailed pace, or extend it by giving Minnesota detective Marge another quirky case to sleuth. Hawley did none of that. <em>Fargo</em> the series carries the Coens&#8217; tones and structures <strong>forward</strong> into new contexts: new decades (1979, 2006, 1950, 2010, 2019), new casts, new crimes, new moral quandaries. Hawley essentially treats the Coens&#8217; entire filmography as the source&#8212;<em>Miller&#8217;s Crossing</em>, <em>No Country</em>, <em>A Serious Man</em>, <em>Burn After Reading</em>&#8212;and lets that whole sensibility propagate through different American eras.</p><p>What&#8217;s striking is that Hawley is explicit about what each season is <em>for</em>. At Canneseries he described Season 2 as the death of the family business and the rise of corporate America; Season 3 as a deconstruction of &#8220;this is a true story&#8221; in the era of alternative facts; Season 5 as a meditation on whether the <em>Yellowstone</em> worldview&#8211;the idea that a man simply knows in his bones what&#8217;s right&#8211;is heroism or villainy. Underneath all of these thematic variations, he said, both <em>Fargo </em>the movie and <em>Fargo</em> the TV show have always been about the battle between decency and cynicism, and in America right now decency isn&#8217;t winning.</p><p>This is franchise expansion by <strong>translation forward to apply prior themes, tones, and tropes to new stories with new characters in different time periods</strong>. The franchise gets bigger not because more of the same thing happened, but because the sensibility proved portable, and because each iteration carries a fresh diagnostic claim about the country it&#8217;s depicting. Which, incidentally, is exactly what the <em>Atlantic</em> billionaire essay is doing in prose form. Hawley has a thesis about consequence and decency, and <em>Fargo</em> is the long-running fictional laboratory for it.</p><h4>DOWN: <em>Legion</em></h4><p>Hawley&#8217;s approach to <em>Legion</em> was a very different maneuver. When faced with the opportunity to do a television show based in the expansive X-Men universe, Hawley monomaniacally drilled <strong>down</strong> into one character: David Haller, a psychically gifted, schizophrenia-diagnosed mutant from a C-list corner of the comics. The result was three seasons of unapologetically weirdo television: bizarre Bollywood dance numbers staged inside psychic battles, silent-film interludes, rap-battle telepathy, a season-long time loop, a villain living inside the protagonist&#8217;s skull. He went deep&#8211;into the character, into the psyche, into his own artistic preoccupations&#8211;and instead of just pulling more of the same from that mineshaft, Hawley excavated something uniquely his.</p><p>His Canneseries pitch for the show was &#8220;What if <em>Breaking Bad</em> was about Walter White becoming a supervillain? I found Professor X&#8217;s son, who&#8217;s mentally ill, basically. He has these powers but he&#8217;s not sure whether they&#8217;re real. And if he doesn&#8217;t know, then that&#8217;s the show.&#8221; He made a deeply philosophical and deeply freaky show all about the instability between what&#8217;s real and what&#8217;s in your head. You can&#8217;t do that shit with Wolverine.</p><p><strong>Expansion through depth of specificity in exploring weird dark corners</strong>. The franchise gets bigger because one narrow slice got unimaginably deeper.</p><h4>OUT: <em>Alien: Earth</em></h4><p><em>Alien: Earth</em>&#8212;FX&#8217;s biggest streaming premiere ever, 94% on Rotten Tomatoes&#8212;is Hawley doing the opposite of <em>Legion</em>. Instead of drilling down on the Xenomorph, he <strong>broadens</strong> the world horizontally. The franchise has spent 45 years on one creature (the titular alien), one megacorporation (Weyland-Yutani), and one form of AI (synthetic persons). Hawley&#8217;s 2120 has five corporations (Prodigy, Weyland-Yutani, Lynch, Dynamic, Threshold), three categories of artificial person (synthetics, cyborgs, and the new hybrid synthetics with uploaded human consciousnesses), and a whole menagerie of horrific <a href="https://nerdist.com/article/new-alien-species-from-alien-earth/">new alien species</a> beyond the Xenomorph (blood ticks, octopus eyes, the orchid and the fly nest).</p><p>Hawley also broadened the franchise&#8217;s applicability to <em>now</em> by literally bringing it home to Earth, and using our current anxieties about AI, billionaires, and infectious, out-of-control biological processes against us.</p><p>This is <strong>expansion by horizontal world-thickening</strong>. The franchise gets bigger thanks to Hawley populating it with more, and more thematically rich, versions of what made us love the series in the first place. The title <em>Alien</em> always had a double-meaning, both a noun and an adjective. Hawley ran with the adjective and broadly applied it to new creatures, new minds, new corporate entities, new fears of the unknown.</p><h4>Forward, Down, and Out</h4><p>Hawley has found a way to rely on pre-existing storyworlds without exhausting them, a way to diagnose a property for what it most needs from him&#8211;and what he most needs from it&#8211;and picking a new direction. He doesn&#8217;t choose the linear-thinker path of retelling the story (the Die Hard 2 model), or extending the plot with the same characters (the Star Wars model), or telling a new adventure of the same protagonist in the same style (the Raiders model). He&#8217;s the kind of showrunner who tellingly doesn&#8217;t rewatch the originals he is adapting before he starts. He works from the <em>emotion</em> he remembers, then asks how to recreate it but in a completely new way.</p><p>This is why Hawley&#8217;s particular skill is so uniquely valuable in this media environment. What&#8217;s scarce in this landscape isn&#8217;t content; it&#8217;s authorial judgment about which direction to grow a property, and why. A generative model can produce infinite Fargo-flavored dialogue, infinite Legion-style freakouts, infinite Xenomorph variants. It cannot decide that <em>this</em> IP needs forward-translation and <em>that</em> one needs vertical compression and <em>that</em> one needs horizontal thickening. It can&#8217;t preserve and expand on the <em>feeling </em>of a storyworld so that the brand is still recognizable to audiences but the form it takes is novel enough to intrigue audiences. The diagnosis is the craft.</p><p>As <a href="https://variety.com/2026/film/news/steven-spielberg-hollywood-original-movies-disclosure-day-cinemacon-1236722625/">Steven Spielberg</a> argued at Cinemacon a few weeks ago, Hollywood needs to invest in more original IP rather than trying to squeeze every last drop out of old IP, otherwise it will creatively &#8220;run out of gas.&#8221; That&#8217;s certainly true. But if you&#8217;re going to invest in legacy IP, and want to do so in a way that is additive rather than exploitative, there&#8217;s no one better than 3D franchise expander Noah Hawley. IP extensions will necessarily continue to be a dominant form in the mediasphere as studios compete with a flood of free online content, but hopefully Hawley&#8217;s uniquely creative approach will help inspire the next generation of franchise adapters to think in more than one dimension.</p><h2>Model Hypocrisy: Unconsented Data Scraping for Me But Not for Thee?</h2><p>In the last issue of <em>Converger</em> I highlighted the <a href="https://kevinbankston.substack.com/i/194807541/happy-100th-ai-lawsuit-to-those-who-celebrate">raft of copyright lawsuits</a> against AI labs currently being litigated in the US and around the world, including challenges to their use of masses of copyrighted expression scraped from the internet to train their models. The labs have consistently argued that using that data for training is defensible as a transformative &#8220;fair use&#8221; under copyright law, and the White House in its <a href="https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf">AI Action Plan</a> made clear it agreed.</p><p>However, the White House&#8217;s Office of Science and Technology Policy (OSTP) just released on April 23rd&#8212;based on apparent pressure from the labs themselves&#8212;a saber-rattling <a href="https://www.whitehouse.gov/wp-content/uploads/2026/04/NSTM-4.pdf">memo</a> that is directly and hypocritically in tension with that fair use position. Specifically, the memo condemned distillation &#8220;attacks&#8221; (i.e., automated querying of model outputs for the purpose of training new models) by Chinese AI labs against US models, after <a href="https://restofworld.org/2026/openai-deepseek-distillation-dispute-us-china/">OpenAI</a>, <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">Anthropic</a>, and <a href="https://cloud.google.com/blog/topics/threat-intelligence/distillation-experimentation-integration-ai-adversarial-use">Google DeepMind</a> all individually complained of the same in February. The US labs variously complained about these &#8220;cyber-attacks&#8221; as &#8220;free-riding&#8221; &#8220;intellectual property theft&#8221; of &#8220;proprietary&#8221; data, and the White House followed suit:</p><blockquote><p>[T]he United States government has information indicating that foreign entities, principally based in China, are engaged in deliberate, industrial-scale campaigns to distill U.S. frontier AI systems. Leveraging tens of thousands of proxy accounts to evade detection and using jailbreaking techniques to expose proprietary information, these coordinated campaigns systematically extract capabilities from American AI models, exploiting American expertise and innovation.</p></blockquote><p>I am mindful that US labs are in a technical race with Chinese labs that has significant economic and national-security implications, and if the US government officially complaining about distillation helps dissuade Chinese action here (it won&#8217;t) then please, go ahead and rattle that saber. But these complaints are broadly and indeed hypocritically in tension with the positions of the AI labs and the US government in other contexts.</p><p>First and most obviously: US AI labs trained their models based on &#8220;industrial-scale&#8221; unconsented copying of other people&#8217;s copyrighted expression, and for them to argue that was fair use while claiming that training on <em>their</em> data is not a fair use is wholly contradictory. Indeed, there are at least four arguments why model distillation is more likely to benefit from the defense of fair use than training on indiscriminately-scraped copyrighted works, or may not even implicate copyright law at all.</p><p><strong>One: </strong>There is a long line of cases standing for the proposition that studying outputs from a technology to reverse engineer it is a fair use, including <em><a href="https://law.justia.com/cases/federal/appellate-courts/F2/977/1510/305345/">Sega Enterprises Ltd. v. Accolade, Inc.</a></em><a href="https://law.justia.com/cases/federal/appellate-courts/F2/977/1510/305345/">, 977 F.2d 1510 (9th Cir. 1992)</a> (disassembling Sega Genesis game code to make compatible third-party games was fair use), <em><a href="https://law.justia.com/cases/federal/appellate-courts/F2/975/832/163650/">Atari Games Corp. v. Nintendo of America Inc.</a></em><a href="https://law.justia.com/cases/federal/appellate-courts/F2/975/832/163650/">, 975 F.2d 832 (Fed. Cir. 1992)</a> ( &#8220;reverse engineering, untainted by the purloined copy of the 10NES program and necessary to understand 10NES, is a fair use&#8221;&#8212;Atari lost on its specific facts only because it had obtained Nintendo&#8217;s source code from the Copyright Office by lying about pending litigation); and <em><a href="https://caselaw.findlaw.com/court/us-9th-circuit/1452245.html">Sony Computer Entertainment, Inc. v. Connectix Corp.</a></em><a href="https://caselaw.findlaw.com/court/us-9th-circuit/1452245.html">, 203 F.3d 596 (9th Cir. 2000)</a> (intermediate copying of Sony&#8217;s PlayStation BIOS during the development of an emulator was protected fair use, even where the resulting product directly competed with Sony&#8217;s console).</p><p>These cases together stand for the proposition that observing what a system does&#8212;including copying its outputs&#8212;to build a competing system is a fair use, particularly where the underlying functional elements are not themselves copyrightable subject matter. Notably, the underlying models themselves, which are complex functional mathematical objects created by algorithmic processes, may not be copyrightable under current law at all.</p><p><strong>Two:</strong> At least in the opinion of the Copyright Office, generative outputs are not copyrightable for lack of a human author, meaning a fair-use defense would not even be necessary since the supposedly &#8220;proprietary&#8221; outputs are not copyright-protected. <em>See</em> U.S. Copyright Office, <em><a href="https://www.copyright.gov/ai/Copyright-and-Artificial-Intelligence-Part-2-Copyrightability-Report.pdf">Copyright and Artificial Intelligence, Part 2: Copyrightability</a></em> (Jan. 29, 2025) (Fully AI-generated outputs generated in response to a human input or prompt lack human authorship and are therefore not copyrightable).</p><p><strong>Three:</strong> to the extent there is a copyright interest in outputs, the labs have typically assigned that interest to the users eliciting the outputs, and disclaimed their own. OpenAI&#8217;s<a href="https://openai.com/policies/row-terms-of-use/"> Terms of Use</a> provide that &#8220;you... own the Output&#8221; and &#8220;we hereby assign to you all our right, title, and interest, <em>if any</em>, in and to Output&#8221; (emphasis mine&#8212;the &#8220;if any&#8221; is the lab itself hedging that there may be no copyright interest there at all). Anthropic&#8217;s and Google&#8217;s consumer terms are<a href="https://www.termsfeed.com/blog/ai-output-licensing/"> substantially similar</a>. The labs have done this in part for self-serving liability reasons&#8212;in the context of potential infringement claims, they want to make clear that the outputs are <em>your</em> creation, not theirs&#8212;but the doctrinal consequence cuts the other way: a lab that disclaims ownership of an output cannot, in the next breath, claim that output as a proprietary asset that competing researchers must not be allowed to learn from.</p><p><strong>Four and finally:</strong> to the extent the labs argue that distillation is prohibited by their terms of service, a lack of a copyright interest in their models or their outputs would call the legality of that prohibition <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5049562">into question</a> (Mark Lemley and Peter Henderson argue that contractual restrictions on the use of AI outputs may be preempted or unenforceable where there is no underlying intellectual property interest to protect). Intellectual property only provides certain limited rights, not a general ability to dictate through contract how information and expression are used. Courts are quite unlikely to consider any outputs to be trade secrets if anyone can elicit them from a model that&#8217;s been exposed to the public; if they also aren&#8217;t copyrighted, there isn&#8217;t really another relevant type of IP law to rely on.</p><p>Indeed, the weakness of any legal argument against model distillation&#8211;and its contradicting of the labs&#8217; own arguments in other cases&#8211;is perhaps best indicated by the fact that they haven&#8217;t sued anyone over it yet. And notably, the White House too fails to articulate any reason why these &#8220;attacks&#8221; are illegal.</p><p>Beyond not making sense as a matter of IP law, the White House&#8217;s position also contradicts its AI Action Plan&#8217;s dedication to fostering the growth of open-source AI models as a source of competitive innovation that isn&#8217;t centralized in the hands of a few large AI labs. Both <a href="https://x.com/deanwball/status/2047468532259045797">Dean Ball</a>&#8212;the AI Action Plan&#8217;s primary staff drafter while at OSTP&#8212;and open-source AI researcher <a href="https://x.com/natolambert/status/2047454390601306207">Nathan Lambert</a> of the Allen Institute for AI (Ai2) have discussed at length on Twitter how distillation is a common technique, including amongst academics, for testing and improving the AI state of the art and developing open-source models including smaller models that can run locally. That open-source ecosystem benefits all AI advancement, not just in China but here and around the world.</p><p>Nor is the use of distillation limited to the open ecosystem: for example, Elon Musk <a href="https://techcrunch.com/2026/04/30/elon-musk-testifies-that-xai-trained-grok-on-openai-models/">admitted</a> on the stand last Thursday in his lawsuit against OpenAI that his company xAI has &#8220;partly&#8221; used distilled outputs from OpenAI models to help train Grok, calling it standard industry practice. The other labs have likely done the same to each other at times. As expressed by <a href="https://x.com/ClementDelangue/status/2050013015680995631">Clem Delangue</a>, the CEO of open-source AI platform Hugging Face, all of this vibes a bit like the labs &#8220;pulling the ladder&#8221; up behind them: </p><blockquote><p>All labs trained their models by distilling (at the very least distilling the web) which allowed them to become the fastest growing businesses in the history of humanity and now that they have armies of lawyers and lobbyists, they are trying to prevent others from doing the same thing.</p></blockquote><p>For all these reasons, the complaints from the major labs and the US government about distillation by foreign competitors ring hollow. There is also the question, raised by Dean Ball, of whether the US government even &#8220;actually adds value in practice on the issue of combatting distillation, and whether that addition of value is larger than the risks associated with the USG paying close attention to the minutiae of frontier AI development.&#8221; This question is all the more salient now that the White House is <a href="https://www.nytimes.com/2026/05/04/technology/trump-ai-models.html">reportedly</a> considering prior restraints on the release of frontier models by US labs, a particularly heavy-handed and constitutionally dubious intervention in the name of security.</p><p>Perhaps, if the labs want to combat distillation more aggressively, they should simply invest more money in their own technical countermeasures and leave the rest of us out of it. Or they can sue someone, if they don&#8217;t think it will completely undermine their own fair use defenses in the scraping cases against them. But I wouldn&#8217;t hold my breath.</p><h2>A Hollywood AI Pipeline Built on Chinese Models?</h2><p>Following up on my piece in the first issue about <a href="https://kevinbankston.substack.com/p/ed2d73d3-f032-4b23-b34c-d354b6c11a33">the post-Sora video gen landscape</a>, fresh news has validated my take: Netflix is investing heavily in both open and proprietary AI video gen tools, and traditional studios that don&#8217;t build their own tech stack will soon be at a distinct disadvantage.</p><p>Netflix&#8212;along with its subsidiary Eyeline Labs&#8212;followed up last month&#8217;s open-source release of the VOID model by dropping <a href="https://eyeline-labs.github.io/Vista4D/">Vista4D, a &#8220;video reshooting framework&#8221;</a> on GitHub. The tool takes previously shot video input and lets you change camera angle, add or modify camera movement, or dynamically expand the scene.</p><p>Like both VOID and the InterPositive tools that Netflix acquired from Ben Affleck, these new tools are focused on modifying already-shot footage rather than generating video from scratch with text prompts. Unlike those other tools, though, this one is built on top of <a href="https://huggingface.co/Wan-AI/Wan2.1-T2V-14B">Wan2.1-T2V-14B</a>, a 14-billion-parameter text-to-video diffusion model open-sourced by Chinese tech giant Alibaba.</p><p>Netflix building on top of a Chinese video model is emblematic of a post-Sora technology trend identified last week by <em><a href="https://www.wsj.com/tech/ai/china-ai-video-generation-f882dccf?mod=e2twd">The Wall Street Journal</a></em>: China is currently whooping American labs on the video generation front, both in adoption of open models like Wan and proprietary models like Kuaishou&#8217;s Kling&#8212;even as some Western labs (<a href="https://techcrunch.com/2026/03/29/soras-shutdown-could-be-a-reality-check-moment-for-ai-video/">per TechCrunch</a>) retreat from consumer video generation entirely.</p><p>As the <em>WSJ</em> reports, Chinese labs took seven of the top 10 spots for video-generation models last month on rating platform Artificial Analysis, including the widely popular (and in Hollywood, reviled) Seedance 2.0 from Bytedance and the new benchmark-smashing HappyHorse model from Alibaba.</p><p>And in the <em>Ankler</em>, Erik Barmack <a href="https://theankler.com/grok-kling-runway-ai-videos-future/">sketches a three-vertical landscape</a>: social generation, led by xAI&#8217;s Grok; professional short-form video, led by Kling; and Runway for Hollywood pre- and post-production. Then again, Kling is popular in Hollywood too: as the <em>WSJ</em> points out, it&#8217;s Kuaishou&#8217;s model that was used by <em>House of David</em>&#8216;s producers for their virtual backlot, and presumably that&#8217;s what they&#8217;re using for their upcoming <em>Moses</em> series (discussed in our last issue).</p><p>And Asia&#8217;s not just leading on video models. While Hollywood debates whether AI belongs on a film set at all, Asian producers are racing ahead to integrate it into the production pipeline itself&#8211;slashing costs and compressing timelines on both feature films and microdramas.</p><p>In China, <em><a href="https://www.economist.com/china/2026/04/09/ai-micro-dramas-are-shaking-up-chinese-entertainment">The Economist</a></em><a href="https://www.economist.com/china/2026/04/09/ai-micro-dramas-are-shaking-up-chinese-entertainment"> reports</a> that AI-animation tools have cut microdrama production costs by up to 90%, with live-action microdrama production dropping 80% in some regions;<em> <a href="https://www.nytimes.com/2026/05/03/world/asia/china-microdrama-ai-backlash.html">The New York Times similarly reports</a></em> on how AI is both fueling and disrupting the Chinese microdrama space.</p><p>In India, <em><a href="https://www.hollywoodreporter.com/business/digital/india-ai-filmmaking-1236548136/">The Hollywood Reporter</a></em><a href="https://www.hollywoodreporter.com/business/digital/india-ai-filmmaking-1236548136/"> detailed last week</a> what it called &#8220;the world&#8217;s most consequential live experiment in AI filmmaking&#8221;: roughly 80% of Indian films now use AI extensively in pre-visualization, AI-driven dubbing is rolling out at platform scale across a dozen languages, and Mumbai studios are producing AI-native features in six to twelve months versus the two-to-three years for traditional animation.</p><p>These two trends are converging faster than anyone expected: the center of gravity in video generation is moving to Chinese models, and the center of gravity in AI-integrated production is moving to Asia more broadly. Netflix is the one American studio racing ahead. The question is which other studios&#8212;or American AI labs&#8212;are even in the race.</p><h2>How <em>Not </em>To Integrate AI Into Newsrooms: McClatchy&#8217;s AI-Driven Byline Blues and OpenAI&#8217;s Slop News Site</h2><p>Moving from the artist&#8217;s studio to the newsroom: last issue we talked about the controversy over when and how to integrate AI into reporting, with some journalists choosing to leverage AI tools and others resisting. Since then we&#8217;ve seen a couple new lessons in how <em>not</em> to use AI in newsrooms.</p><p>First lesson is from media company McClatchy, operator of thirty newspapers across the country, which as first reported in multiple stories by <em><a href="https://www.thewrap.com/media-platforms/journalism/mcclatchy-ai-tool-revolt-sacramento-bee-miami-herald-charlotte-observer/">The</a> <a href="https://www.thewrap.com/media-platforms/journalism/mcclatchy-content-scaling-agents-roiling-newsrooms/">Wrap</a></em> recently introduced a new AI-driven &#8220;Content Scaling Agent&#8221; into all its newsrooms. &#8220;More stories, more inventory&#8221; was management&#8217;s hope and demand as they launched the tool to reporters with a groan-worthy <em><a href="http://x.com/CorbinBolies/status/2046593147346551066">Star Wars</a></em><a href="http://x.com/CorbinBolies/status/2046593147346551066">-style intro crawl</a>.</p><p>It would have been controversial enough just for McClatchy to try to force reporters to use AI to rewrite and repackage their original stories for &#8220;new audiences, angles and entry points.&#8221; But as also reported in the <em><a href="https://www.nytimes.com/2026/05/01/business/media/mcclatchy-ai-newsroom-byline-strike.html">The New York Times</a></em> last week, McClatchy truly stirred up the hornet&#8217;s nest when it said that it would use reporters&#8217; bylines on AI-written stories adapted from their work even against their objections, in order to help get better Google rankings. &#8220;If they don&#8217;t have the ability in their contract to remove their byline, we&#8217;re going to use their name,&#8221; said management, leading to complaints from several of the newsrooms&#8217; unions and a joint letter campaign by writers to withhold their bylines.</p><p>As Ariane Lange, an investigative reporter at the <em>Sacramento Bee</em> and the vice chair of its union, told the <a href="https://www.cjr.org/analysis/fighting-the-machine-contracts-artificial-intelligence-ai-use-bylines.php#:~:text=And%20lately%2C%20employees%20at%20McClatchy%20newsrooms%20have,scaling%20agent%2C%E2%80%9D%20an%20AI%20tool%20powered%20by">Columbia Journalism Review</a>,</p><blockquote><p>I&#8217;ve covered traffic deaths in the city of Sacramento since 2024, and I have talked to many families of people who have been killed in crashes, and that&#8217;s a very vulnerable moment. I&#8217;m assuring them they can trust me, but I also have to explain that my employer might feed their story to a chatbot and spit it back out as five key takeaways. That&#8217;s revolting to me.</p></blockquote><p>In an equally revolting development first reported by <a href="https://www.modelrepublic.org/articles/the-reporters-at-this-news-site-are-ai-bots.-openai%E2%80%99s-super-pac-appears-to-be-using-it-to-advance-its-political-agenda">Model Republic</a>, a PR firm working for the AI industry was revealed in late April to have been running a completely fake AI-driven news site with stories attacking AI regulation and the industry&#8217;s critics. The discovery came when one of those critics received a fishy inquiry from a purported journalist from the site, <a href="https://acutuswire.com/">The Wire by Acutus</a>, that when run through the AI detector Pangram was ID&#8217;d as 100% AI generated.</p><p>On further investigation it appeared The Wire didn&#8217;t have any human journalists at all, and that it was linked to PR firm Novus working for the anti-AI-regulation super PAC Leading the Future that&#8217;s funded by OpenAI&#8217;s president Greg Brockman and venture firm a16z. When news of the bot-driven slopaganda site leaked, the PAC essentially <a href="https://x.com/LeadingFutureAI/status/2047747501738856575">confirmed</a> it was created by their PR firm but without their knowledge, and said they were cutting those ties. Paired with recent news of the same PAC funding <a href="https://www.wired.com/story/super-pac-backed-by-openai-and-palantir-is-paying-tiktok-influencers-to-fear-monger-about-china/">anti-AI-regulation astroturf campaigns</a> amongst TikTok creators, this story certainly paints an unsavory picture of that PAC, their backers, and their tactics.</p><p>Thankfully, AI efforts like McClatchy&#8217;s that disempower rather than support journalists are receiving <a href="https://reutersinstitute.politics.ox.ac.uk/news/nailing-jell-o-wall-why-unions-are-struggling-protect-journalists-rights-age-ai?utm_source=chatgpt.com">sustained pushback</a> from newsroom unions in their negotiations, from the <a href="https://www.thewrap.com/media-platforms/journalism/new-york-times-union-contract-negotiations-ai/">New York Times</a> to <a href="https://www.thewrap.com/media-platforms/journalism/propublica-union-walkout-ai-contract-negotiations/">ProPublica</a> to <a href="https://www.thewrap.com/media-platforms/journalism/cbs-news-union-deal-ai/">CBS News</a>. No news yet, however, on whether the Wire&#8217;s robot reporters are also organizing to improve their own working conditions.</p><h2>Hannah Einbinder Swirlie Watch: Who&#8217;s Getting Flushed for Using AI This Week?</h2><p>Last issue we <a href="https://kevinbankston.substack.com/i/194807541/steven-soderbergh-volunteers-to-get-a-swirlie-from-hannah-einbinder">highlighted</a> how <em>Hacks</em> star comedian Hannah Einbinder went off on AI creators, calling them losers and saying she wanted to stick their heads in a toilet and flush (what any high-school bully knows as a &#8220;swirlie&#8221;). In the past few weeks, several new media figures have risked Hannah&#8217;s ire, and more will certainly volunteer for swirlies in the coming months, so I figure that Swirlie Watch will become a semi-regular feature here.</p><p>Who should be on guard against Hannah&#8217;s swirlie vengeance this week?</p><ul><li><p><a href="https://deadline.com/2026/04/alex-proyas-teams-with-ai-producer-ex-machina-heaven-1236869390/?utm_source=chatgpt.com">Director Alex Proyas</a> (of <em>The Crow</em> and <em>I, Robot</em>) announced that he&#8217;d be working with AI studio Ex Machina to generate his next feature, a digital-afterlife sci-fi satire called <em>Heaven</em>.</p></li><li><p><a href="https://deadline.com/2026/04/roger-avary-direct-biblical-paradise-lost-ai-production-1236875688/">Writer-Director Roger Avary</a> (co-writer and producer of <em>Pulp Fiction</em>), also in partnership with Ex Machina, announced he&#8217;ll be adapting Milton&#8217;s <em>Paradise Lost </em>using generative AI.</p></li><li><p><a href="https://au.variety.com/2026/film/news/shawn-levy-ai-become-essential-tool-moviemaking-but-he-hasnt-incorporated-meaningful-way-yet-35741/">Director Shawn Levy</a> (<em>Deadpool &amp; Wolverine</em>, the upcoming <em>Star Wars: Starfighter</em>) says he hasn&#8217;t &#8220;meaningfully&#8221; used gen AI yet in his own storytelling&#8211;so maybe he just gets toilet-dunked rather than fully swirled?&#8211;&#8220;but I have no doubt that in the course of my career we will see its integration&#8230; I think it&#8217;s going to be essential.&#8221; Nope, that&#8217;s a full swirlie right there.</p></li><li><p><a href="https://www.thewrap.com/industry-news/industry-trends/reese-witherspoon-responds-ai-backlash/">Actor-Producer Reese Witherspoon</a> doubled-down on her comments urging women to learn AI, while making clear &#8220;I don&#8217;t believe computers should replace humanity&#8221; (phew, you had us worried there!).</p></li><li><p><a href="https://deadline.com/2026/05/steven-soderbergh-talks-ai-john-lennon-doc-1236876040/">Director Steven Soderbergh</a> tripled-down on his comments that we reported on last issue with a much more fulsome explanation of the AI use in his upcoming John Lennon and Yoko Ono documentary, in partnership with Meta, defending that choice while also being fully transparent about it.</p></li><li><p><a href="https://www.thewrap.com/creative-content/movies/ai-doug-liman-movie-reaction-jobs-fear-commentary/">Producer Ryan Kavanaugh</a> had the temerity to defend the virtually-produced Doug Liman thriller, <em>Bitcoin: Killing Satoshi</em>, raising the same possibility I articulated last issue: that virtual production might actually be a way to bring more movie productions and more industry jobs back to LA.</p></li><li><p><a href="https://x.com/diplo/status/2044034908985946449">DJ and music producer Diplo</a> said in a podcast interview and then reiterated on Twitter that &#8220;there&#8217;s no fighting AI&#8221; and &#8220;if you are a creative you need to adapt or just like give up and become an uber driver until everyone has a waymo.&#8221; That definitely warrants multiple flushes.</p></li><li><p><a href="https://www.theguardian.com/music/2026/apr/21/jean-michel-jarre-music-film-industries-embrace-ai">French electronic music pioneer Jean Michel-Jarre</a> complained of the conservatism of creative industries &#8220;freaking out&#8221; about AI, predicting that artists would use AI &#8220;to create the cinema of tomorrow, the hip-hop of tomorrow, the techno of tomorrow, the rock&#8217;n&#8217;roll of tomorrow.&#8221; He says &#8220;[w]e should never be afraid of technology.&#8221; But <em>he</em> should be afraid of Hannah Einbinder.</p></li><li><p><a href="https://variety.com/2026/digital/columns/openai-ceo-sam-altman-ai-in-hollywood-will-get-people-to-care-more-about-human-creators-not-less-1236726413/">OpenAI CEO Sam Altman</a> pushed back on Hollywood concerns about AI replacing humans with some unconvincing pro-human pablum: &#8220;I think people really care about other people&#8230;. I think people really care about the human beings behind the stories and the art and the creative work that matters so much, so my instinct is it&#8217;s going to go the other way and people will care more about humans and more about human creators in the future, not less.&#8221; Problem solved!</p></li><li><p><a href="https://www.hollywoodreporter.com/movies/movie-news/nyu-film-school-runway-ai-tools-spike-lee-1236560902/">NYU Film School</a>&#8217;s administration and faculty earned a collective swirlie by partnering with Runway to put AI credits and training in the hands of its students for free (well, plus NYU&#8217;s exorbitant tuition). Looks like <a href="https://theankler.com/usc-to-nyu-ais-stealth-film-school-takeover-has-begun/">USC&#8217;s film school and the Sundance Institute</a> may also have earned a dunking for getting in bed with Adobe and Google.</p></li></ul><p>Get to work on those swirlies, Hannah! You have a long target list this week. Everyone else, be on guard if entering a public bathroom in Hollywood or NYC.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading CONVERGER! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h1>FRAGMENTS</h1><h2>Creator, Trademark Thyself</h2><p><a href="https://www.hollywoodreporter.com/business/business-news/taylor-swift-filing-for-trademarks-combat-ai-misuse-can-it-work-1236578397/">Taylor Swift</a> is following <a href="https://www.wsj.com/tech/ai/matthew-mcconaughey-trademarks-himself-to-fight-ai-misuse-8ffe76a9">Matthew McConaughey&#8217;s</a> lead and filing to trademark her voice and image in an attempt to help combat misuse of both by AI. The actor&#8217;s <a href="https://variety.com/2026/film/news/matthew-mcconaughey-timothee-chalamet-ai-actors-oscars-1236667017/">advice</a>: &#8220;So I say: Own yourself. Voice, likeness, et cetera. Trademark it. Whatever you gotta do, so when [AI] comes, no one can steal you.&#8221; Well alright alright alright.</p><h2>ChatGPT Images 2.0 Can Now Fake Your Doctor&#8217;s Note</h2><p>The new image model is, as highlighted by <a href="https://techcrunch.com/2026/04/21/chatgpts-new-images-2-0-model-is-surprisingly-good-at-generating-text/">TechCrunch</a>, surprisingly good at rendering coherent and readable text outputs, leading to <a href="https://x.com/mark_k/status/2046501686324187332">all</a> <a href="https://x.com/yanatweets/status/2046678255215010198">kinds</a> <a href="https://x.com/nicdunz/status/2046694358410309980">of</a> <a href="https://x.com/DennisonBertram/status/2048413815675539816">novel</a> <a href="https://x.com/TheAtlantic/status/2050577306008113556">use</a> <a href="https://x.com/immasiddx/status/2046969370631717326">cases</a> popping up on Twitter.</p><h2>Affleck Tops List of Most Powerful AI Players in Hollywood</h2><p><em>The Hollywood Reporter</em> published its list of <a href="https://www.hollywoodreporter.com/lists/most-powerful-people-ai-2026/">the 25 most powerful people shaping the future of AI</a> in Hollywood, and at the top was Ben Affleck whose AI startup sale to Netflix has helped prompt a vibe shift in the creative world around using AI tools in filmmaking. Or, as one commentator put it to <a href="https://www.vanityfair.com/hollywood/story/ben-affleck-netflix-celebrities-using-ai">Vanity Fair</a>: having a prominent creator like Ben Affleck pushing AI &#8220;helps get filmmakers to feel more comfortable [using it]&#8212;and that&#8217;s worth the $600 million&#8221; that Netflix paid him. For my part I&#8217;m very supportive of the creative use of AI but &#8220;he got paid hundreds of millions to use his celebrity to normalize AI&#8221; seems a bit gross, and as reported in the <a href="https://kevinbankston.substack.com/i/194807541/pay-no-attention-to-the-foundation-model-behind-ben-afflecks-curtain">last issue</a>, Affleck and Netflix also haven&#8217;t been very forthcoming about the technology behind the startup&#8217;s product nor about the copyrighted works behind that technology. But being the most powerful isn&#8217;t always pretty!</p><h2>New Creative AI Integrations, Integrating AI Creatively</h2><p>The last couple weeks saw some big releases in the realm of integrating creative tools into chatbots and chatbots into creative tools.</p><p>On April 23rd, a16z and Union Square Ventures-backed startup <a href="http://glif.app">Glif</a> came out of beta with the <a href="https://x.com/fabianstelzer/status/2047359946702880920">release of Glif v2</a>, pitched as &#8220;it&#8217;s like Claude Code for AI videos&#8221;: a &#8220;creative super agent&#8221; chatbot interface that integrates with virtually every available video and audio gen model for a unified creative dashboard.</p><p>Five days later the makers of actual Claude announced <a href="https://www.anthropic.com/news/claude-for-creative-work">Anthropic&#8217;s own integrated creative offering</a> with a raft of new connectors between Claude and a wide range of creative applications: the entire Adobe Cloud suite of tools, Blender, Canva&#8217;s Affinity, Ableton, Splice, SketchUp, and Resolume. This is a clear warning to startups like Glif: when building wrapper interfaces around other people&#8217;s models and services, you risk the model-makers themselves offering the same features directly.</p><p>And, in a story I missed before putting out the previous issue, Google and Avid announced earlier in April the <a href="https://www.latimes.com/entertainment-arts/business/story/2026-04-16/hollywood-editors-get-new-ai-tool-from-google-avid">integration of Gemini AI features into the Avid editing suite</a>, a godsend for editors trying to locate just the right shot in a heaping digital pile of other takes. Harried editors can now just describe in chat the particular visual movements, dialogue or emotional cues they&#8217;re looking for and Gemini will help surface the clip they need.</p><p>Regardless of whether you&#8217;re using AI to generate art or just tweak or edit it, and whether you&#8217;re using a startup&#8217;s toolkit or subscribing to a foundation lab&#8217;s frontier model, creative AI is getting exponentially easier to use by the day, and the trend will likely continue to be toward one creative interface to rule them all.</p><h2>Because Parasocial Relationships with AI Aren&#8217;t Weird Enough</h2><p>The dominant player in vertical comics, Webtoon, has <a href="https://variety.com/2026/gaming/news/webtoon-entertainment-genies-ai-1236729996/">inked a deal</a> with AI avatar tech company Genies to let users chat with their favorite comic characters. Still unclear what the content guardrails or age restrictions around that tech might be but worth noting that Disney has a deal with Webtoon to build <a href="https://bleedingcool.com/comics/disney-marvel-webtoon-to-launch-new-digital-comics-app-and-platform/">Marvel&#8217;s new digital comic platform</a>, so we may end up seeing this tech applied to many more popular characters than just those in your favorite manga.</p><h2>Because Parasocial Relationships with AI Aren&#8217;t Weird Enough, Part Two</h2><p>The New Yorker published a fascinating and disturbing article about TikTok and Instagram influencers using AI to pose as different races and genders; as the author put it on Twitter, <a href="https://x.com/TM_Brown/status/2048027421795037682">&#8220;it gets weird real fast.&#8221;</a> See also this related Wired story on<a href="https://x.com/ejdickson/status/2046588692651815340">fake MAGA girl influencers</a>, and this <a href="https://x.com/andreysuperior/status/2050908800303915020">creepy demo video</a> I found on Twitter of a college dude using AI to pose as a bikini blonde. As the article retweeted by that poster noted, you now never know whether the influencer you&#8217;re watching is made out of four files on a laptop somewhere.</p><h2>The <em>Other</em> Ryan Gosling Astronaut Movie</h2><p>I really enjoyed <em>Project Hail Mary</em>&#8212;go see it on a big screen before it&#8217;s gone!&#8212;which reminded me of another (tragically underappreciated) movie starring Ryan Gosling as an astronaut: Damien Chazelle&#8217;s Neil Armstrong biopic <em>First Man</em>. That earlier movie inspired me several years ago to write an <em>Atlantic</em> article about <a href="https://www.theatlantic.com/science/archive/2018/10/first-man-sci-fi-science-feedback-loop/572821/">the feedback loop between movies and science</a>, using the example of how the orbital mechanics of the Apollo missions were presaged decades prior in Fritz Lang&#8217;s 1929 silent film <em>Woman in the Moon</em>, informed by Hermann Oberth, the father of German rocketry and the first movie science consultant. (You can read several more of my <a href="https://slate.com/author/kevin-bankston">articles about the sci-fi feedback loop</a> over at <em>Slate</em>).</p><h2>Warner Bros. Shareholders Approve Paramount-Skydance Merger Bid</h2><p>Former FTC commissioner Alvaro Bedoya briefly <a href="https://x.com/BedoyaUSA/status/2047332844943348105">highlights</a> ways it might still be stopped; Wade Major of <a href="https://open.substack.com/pub/hollywoodheretic/p/100-days-of-madness-netflix-paramount?r=13l5g&amp;utm_medium=ios">Hollywood Heretic</a> dives deep with nearly 15,000 words on why he thinks the Netflix takeover was always doomed and the Paramount merger will be better for Hollywood. I&#8217;d still prefer no merger at all but it&#8217;s a fascinating piece.</p><h2>Window Treatment: (Almost) All Studios To Let Movies Stay in Theaters Longer</h2><p>Ever since the pandemic, studios have been aggressively shortening theatrical windows for their movies in a short-sighted attempt to shore up their streaming operations. But it looks like the windowing wars may have reached a cease fire as the studios are finally listening to exhibitors, and seeing more clearly how this self-cannibalization has been leaving theatrical money on the table. Based on a slew of recent announcements, longer windows appear to be the <a href="https://www.thewrap.com/creative-content/movies/theatrical-windows-box-office-45-days/">&#8220;new gospel&#8221;</a> in Hollywood: Last Friday Netflix <a href="https://variety.com/2026/film/news/greta-gerwig-narnia-netflix-release-date-1236735796/">announced</a> Greta Gerwig&#8217;s <em>Narnia: The Magician&#8217;s Nephew</em> will get a true wide release in February 2027 with a <a href="https://www.hollywoodreporter.com/movies/movie-news/netflix-narnia-movie-greta-gerwig-release-date-1236581880/">49-day exclusive run</a>&#8212;Netflix&#8217;s first real theatrical window after Ted Sarandos <a href="https://variety.com/2026/tv/news/ted-sarandos-senate-hearing-netflix-warner-bros-deal-1236651030/">pledged</a> 45 days under oath to a Senate antitrust subcommittee in February. It caps a month that saw every major <a href="https://celluloidjunkie.com/2026/04/30/cinemacon-2026-the-studios-make-their-case-for-theatrical/">reaffirm</a> a 45-day floor at CinemaCon: Sony&#8217;s Tom Rothman <a href="https://variety.com/2026/film/news/sony-pictures-boss-cinemacon-urges-fewer-ads-trailers-1236720830/">told</a> exhibitors to enforce it, Paramount&#8217;s David Ellison made a <a href="https://variety.com/2026/film/news/david-ellison-cinemacon-paramount-windows-streaming-1236722998/">surprise appearance</a> to commit to it effective immediately, and Disney <a href="https://www.indiewire.com/news/breaking-news/disney-cinemacon-avengers-doomsday-wild-horse-nine-1235189616/">reiterated</a> its 57-day average. Apple&#8217;s the only studio that hasn&#8217;t yet committed to a longer window, but I&#8217;m guessing its eventual <em>F1</em> sequel will be in theaters for a long, long time.</p><h2>Proving (or Pretending) You Didn&#8217;t Use AI for Your Writing</h2><p>Following up on her <a href="https://www.nytimes.com/2026/03/25/opinion/shy-girl-ai-publishing.html?unlocked_article_code=1.V1A.Bd_F.A9zrlmIe9phc&amp;smid=url-share">New York Times editorial</a> on the AI-driven erosion of trust between authors and readers, novelist Andrea Bartz wrote an <a href="https://andibartz.substack.com/p/sohow-can-you-prove-your-works-not?r=13l5g&amp;utm_medium=ios&amp;triedRedirect=true">in-depth Substack article</a> on the challenges of proving that you didn&#8217;t use AI in your writing, ultimately concluding that the best way to fend off accusations for now is to thoroughly document your writing process. (Notably, Bartz is the lead class representative in <em><a href="https://publicknowledge.org/courts-agree-ai-training-ruled-as-fair-use-in-bartz-v-anthropic-and-kadrey-v-meta/">Bartz v. Anthropic</a></em>, which along with <em>Kadrey v. Meta </em>is the flagship litigation against the tech giants for training their AI models on her and other authors&#8217; novels without permission. Another lawsuit against Meta&#8211;and Mark Zuckerberg personally&#8211;by novelist Scott Turow and five publishers just dropped <a href="https://www.nytimes.com/2026/05/05/books/publishers-turow-meta-zuckerberg-lawsuit-copyright.html">this week</a>.)</p><p>Meanwhile, there&#8217;s a fun new AI tool for those who want to humanify their emails, from Dorm Room Fund tech investor and tinkerer Ben Horwitz: <a href="http://sinceerly.com">Sinceerly.com</a>. Horwitz calls it the <a href="https://x.com/horwitzben/status/2047293550342152680">anti-Grammarly</a>, and it uses AI to mess up your email drafts to make them seem more like something a person would write. There are three settings: the minimal-changes Subtle setting, the intermediate Human setting, and&#8211;with maximum typos and abruptness and minimal grammar compliance&#8211;the CEO setting &#128514;</p><h2>The Actors Guild Deal Has&#8230;Some Sort of AI Protections?</h2><p>Last issue I bemoaned the lack of new protections against AI replacement for members of the Writers Guild of America in their fresh four-year deal with the studios, and pointed to the upcoming SAG-AFTRA negotiations as another opportunity for Hollywood&#8217;s creative class to secure AI concessions. Well, now that actors&#8217; union has also signed a four-year deal with the studios and it has&#8230;some sort of AI &#8220;guardrail&#8221; provisions? Or &#8220;safeguards&#8221; or &#8220;protections&#8221;? But no one <a href="https://deadline.com/2026/05/sag-aftra-studios-reach-new-bigger-deal-amptp-1236879157/">reporting on the deal</a> seems to have any details on what these protections are. So, um, hopefully they&#8217;re good?</p><h2>AI Film Wins at a Traditional Film Festival for First Time; Academy Nixes Awards for AI Writing and Performances</h2><p>Depending on your perspective this is either a travesty or an auspicious first, but: in mid-April, <em>Memory of Princess Mumbi</em>&#8211;a feature film from Swiss-Kenyan director Damien Hauser that substantially used AI to generate its futuristic sci-fi background&#8211;<a href="https://variety.com/2026/film/festivals/memory-of-princess-mumbi-ai-istanbul-festival-1236725769/">won the top prize</a> at the Istanbul Film Festival. As Hauser <a href="https://variety.com/2025/film/global/memory-of-princess-mumbi-damien-hauser-ai-1236499344/">discussed</a> before winning the award, his portrait of a retro-futuristic Africa in the year 2093 would not have been possible without AI, even as he set out to &#8220;make a movie that AI [alone] could never make.&#8221;</p><p>On the other side of the world, the Academy Awards organization released on Friday <a href="https://www.cnet.com/tech/services-and-software/academy-motion-picture-arts-sciences-new-oscars-rules-ai-2026-news/">new rules</a> making clear that AI-generated performers are not eligible for acting awards (so, e.g., no award for <a href="https://variety.com/2026/film/news/val-kilmer-ai-as-deep-as-the-grave-trailer-1236722342/">AI-resurrected Val Kilmer</a>), and only human-authored scripts are eligible for screenwriting awards. The rules notably did <em>not </em>address other uses of AI in regard to other awards&#8211;so, e.g., something like <em>Memory of Princess Mumbi </em>could still be considered for Best Foreign Film.</p><h2>44% Slop and Rising: AI-Generated Music on Deezer</h2><p>Late last month music platform Deezer (which I admittedly had never heard of before this story) <a href="https://techcrunch.com/2026/04/20/deezer-says-44-of-songs-uploaded-to-its-platform-daily-are-ai-generated/">disclosed</a> that 44% of daily uploads&#8211;roughly 75,000 tracks a day, over 2 million a month&#8211;are AI-generated. But consistent with last issue&#8217;s piece on how <a href="https://kevinbankston.substack.com/i/194807541/old-music-beats-new-music-beats-ai-https://kevinbankston.substack.com/i/194807541/old-music-beats-new-music-beats-ai-musicmusic">older music is resonating</a> with streaming listeners much more than new music is, whether AI-generated or not, the AI music on Deezer represents only 1&#8211;3% of streams. And 85% of <em>those</em> streams are flagged as fraudulent and demonetized. The company is now <a href="https://www.musicbusinessworldwide.com/75000-ai-generated-tracks-now-flood-deezer-daily-representing-44-of-all-new-music-uploaded-to-the-platform-says-streamer/">licensing its AI detector</a> to other companies.</p><div><hr></div><p>And that&#8217;s what&#8217;s converging this week! See you next time, and don&#8217;t forget to share if you liked what you read.</p><div><hr></div><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading KCONVERGER! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://converger.kevinbankston.com/p/converger-2-youtubes-concerning-crackdown?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><p></p>]]></content:encoded></item><item><title><![CDATA[CONVERGER #1 — Super-Sized First Edition]]></title><description><![CDATA[Mapping the content singularity where all media collapse into one]]></description><link>https://converger.kevinbankston.com/p/converger-1-super-sized-first-edition</link><guid isPermaLink="false">https://converger.kevinbankston.com/p/converger-1-super-sized-first-edition</guid><dc:creator><![CDATA[Kevin Bankston]]></dc:creator><pubDate>Tue, 21 Apr 2026 15:03:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8acdc050-72fa-4d87-90ab-8e7647dd1c9a_939x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Welcome to CONVERGER, a biweekly newsletter mapping the content singularity where AI and the internet collapse all media into one&#8212;a connective node where emerging technology, policy, culture, futures thinking and storytelling intersect.</p><p>CONVERGER presents news and views from an AI, internet and media policy expert who is pro-innovation but anti-hype, allergic to both AI panic and AI boosterism, and passionate about supporting rather than supplanting human creativity with new technology.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Kevin Bankston's CONVERGER! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>I&#8217;m Kevin Bankston, your host. Since this is the first issue of CONVERGER, it&#8217;s a massively super-sized edition covering the past month.</p><p>Some issues may be heavier on media commentary, others on AI policy, others on personal passions like sci-fi&#8217;s influence on technology (both for good and bad) or the evolving medium and business of comic books in the digital age. You never know what threads might come together in convergence-space!</p><p>Going forward, you can watch me develop newsletter content in real-time on <a href="http://linkedin.com/in/kevinbankston">LinkedIn</a> and the social network formerly known as <a href="https://x.com/KevinBankston">Twitter</a>, and less often on <a href="https://bsky.app/profile/bankston.bsky.social">Bluesky</a> and <a href="https://www.instagram.com/that_kevin_bankston/">Instagram</a>. </p><p>You can also look for my deeper policy-oriented takes on AI governance generally at <a href="https://elicitation.substack.com/">Elicitation</a>, the new Substack from my AI policy day-job colleague Miranda Bogen of the <a href="http://cdt.org">Center for Democracy &amp; Technology</a>&#8217;s AI Governance Lab.</p><p>Now, let&#8217;s converge!</p><div><hr></div><p><strong>TABLE OF CONTENTS</strong></p><p><strong>FEATURES</strong></p><ol><li><p><strong>Happy 100th AI Lawsuit to Those Who Celebrate! </strong><em><strong>(~310 words, 1 minute read)</strong></em></p></li><li><p><strong>Forecasting Four Fraught Futures for the Web in the AI Age </strong><em><strong>(~540 words, 2 minute read)</strong></em></p></li><li><p><strong>Steven Soderbergh Volunteers To Get a Swirlie from Hannah Einbinder </strong><em><strong>(~590 words, 2.5 minute read)</strong></em></p></li><li><p><strong>Hollywood&#8217;s So Angry at AI It Can&#8217;t Spell Straight </strong><em><strong>(~510 words, 2 minute read)</strong></em></p></li><li><p><strong>Pay No Attention to the Foundation Model Behind Ben Affleck&#8217;s Curtain </strong><em><strong>(~1075 words, 4 minute read)</strong></em></p></li><li><p><strong>Director Robert Rodriguez on AI as Creative Multiplier </strong><em><strong>(~445 words, 1.5 minute read)</strong></em></p></li><li><p><strong>Virtual Studios: What&#8217;s the Difference Between </strong><em><strong>Sin City</strong></em><strong> and Generated Cities? </strong><em><strong>(~670 words, 2.5 minute read)</strong></em></p></li><li><p><strong>So Long Sora, Welcome Back Seedance 2.0, Hello to Netflix&#8217;s VOID </strong><em><strong>(~540 words, 2 minute read)</strong></em></p></li><li><p><strong>Writers vs. AI vs. Writers </strong><em><strong>(~590 words, 2.5 minute read)</strong></em></p></li><li><p><strong>A Contract That Helps Protect Comic Artists Against AI Training on Their Work </strong><em><strong>(~360 words 1.5 minute read)</strong></em></p></li></ol><p><strong>FRAGMENTS</strong></p><ol><li><p><strong>WGA Deal on AI: Is That All There Is?</strong></p></li><li><p><strong>News of the Supreme Court&#8217;s Ruling on AI and Copyright Has Been Greatly Exaggerated</strong></p></li><li><p><strong>Macro-Growth in the Micro-Drama Content Pipeline</strong></p></li><li><p><strong>Meanwhile, In the YouTube-to-Theaters Pipeline: </strong><em><strong>Backrooms</strong></em><strong> is Coming</strong></p></li><li><p><strong>Old Music Beats New Music Beats AI Music</strong></p></li><li><p><strong>New Script-Reviewing AI Really Wants Brett Ratner to Direct Your Movie</strong></p></li><li><p><strong>Is Fan-Created Content Supplanting Canonical Content?</strong></p></li><li><p><strong>The Future&#8217;s So Bright I Have To Enter This Contest</strong></p></li><li><p><strong>Big Brother, Generating Slop</strong></p></li><li><p><strong>Webtoon Translations, Digital Comics, Tiny Onions</strong></p></li></ol><div><hr></div><h1><strong>FEATURES</strong></h1><h2><strong>Happy 100th AI Lawsuit to Those Who Celebrate!</strong></h2><p>At the border between old media and new technologies, there are lawsuits&#8212;a lot of lawsuits. And on April 3rd we hit a key milestone: the filing of the <a href="https://chatgptiseatingtheworld.com/2026/04/04/updated-u-s-map-of-copyright-suits-v-ai-companies-apr-3-2026/">100th copyright-related lawsuit</a> against an AI company in the US.</p><p>The honor of bringing the 100th case goes to Ted Entertainment, a YouTube creator that simultaneously filed three cases against OpenAI, Apple, and Amazon, alleging they illegally bypassed YouTube&#8217;s technical protections to scrape videos for AI training.</p><p>Meanwhile, what was already the largest copyright settlement in history&#8212;a minimum of $1.5 billion dollars in the case of <em>Bartz v. Anthropic</em>, brought by a class of authors representing every writer of every one of the nearly half a million works in the &#8220;shadow libraries&#8221; of pirated books that Anthropic trained its models on&#8212;has also hit <a href="https://chatgptiseatingtheworld.com/2026/04/16/adding-to-their-largest-settlement-in-us-copyright-history-bartz-book-authors-class-announces-another-historic-milestone-91-3-participation-of-total-works-in-class-wow/">another milestone</a>. The plaintiffs have reported that the authors of 91.3% of the eligible works have registered their claim for part of the settlement. </p><p>As any lawyer reading this knows, that&#8217;s an insane claims rate; the typical consumer class action lawsuit has a rate closer to 10%. But it looks like the potential for approximately $3000 to each author per work was highly motivating to the class members. And as lawsuits against AI labs on behalf of authors proliferate, there may be more settlements coming down the pike.</p><p>This is all according to <a href="https://chatgptiseatingtheworld.com/">ChatGPT is Eating The World</a>, the most comprehensive newsletter tracking copyright law developments around AI. It&#8217;s run by law professor Edward Lee, who commemorated the 100th lawsuit by debuting a new <a href="https://chatgptiseatingtheworld.com/2026/04/09/todays-launch-of-new-ai-copyright-case-tracker-via-chatgpt-is-eating-the-world/">dashboard</a> for tracking all of them, both in the US and globally. Handy for AI and copyright nerds!</p><h2><strong>Forecasting Four Fraught Futures for the Web in the AI Age</strong></h2><p>Speaking of AI, scraping, and the law: late last month at Georgetown Law School I had the privilege of hosting a day-long event of public and private panels, talks, and roundtable conversations about &#8220;<a href="https://www.law.georgetown.edu/tech-institute/news/public-forum-internet-scraping-and-the-future-of-the-open-web-in-the-internet-age/">Internet Scraping and the Future of the Open Web in the AI Age</a>,&#8221; in my dual capacities as Senior Advisor on AI Governance at the Center for Democracy &amp; Technology and adjunct professor of AI and copyright law.</p><p>The day was focused on the question of how to preserve an open and sustainable internet ecosystem when human traffic is quickly being supplanted by AI-driven bots, and how the results of those hundred lawsuits mentioned above might impact the answer.</p><p>This tension was highlighted just this last week by <a href="https://www.wired.com/story/the-internets-most-powerful-archiving-tool-is-in-mortal-peril/">reporting</a> about how the bots of the Internet Archive, a critically important nonprofit library of internet content, are increasingly being blocked by news websites afraid of their content being scraped by AI companies. That in turn led over a hundred journalists to write an <a href="https://www.fightforthefuture.org/news/2026-04-13-100-journalists-applaud-the-internet-archives-role-in-preserving-the-public-record">open letter</a> defending the importance of the Archive&#8217;s record of internet history to their and other researchers&#8217; work, as well as to the public.</p><p>The issue of scraping and AI raises hard questions with a lot of important perspectives on different sides, so our event certainly didn&#8217;t resolve the problem! But we did bring together all the different communities with a vested interest in the issue for a generative [pun intended] dialogue, including commercial and non-commercial AI labs, web publishers and service providers, content delivery networks, librarians and archivists, scraping companies, and academic and public interest experts.</p><p>My favorite part of the day was one of the private expert sessions where we used a common foresight tool&#8212;<a href="https://foresight.unglobalpulse.net/blog/tools/2x2-matrix-scenario-building/">the 2x2 scenarios matrix</a>&#8212;as the basis for a breakout group exercise considering four very different possible futures for the internet based on the different ways companies, the courts, and lawmakers might approach the issue:<br><br><strong>The Wild West:</strong> Unconsented bot scraping is both legal and technically easy and the internet is completely overrun by automated traffic.<br><strong>The Hollow Victory:</strong> Web publishers beat the bots in court but lose the technical fight, so sites continue to be pummeled by abusive scraping from bad actors that are outside the reach of the law.<br><strong>The Gated Web: </strong>Both legal protections and technical protections against bots are strong, leading to an internet of pay-per-scrape content fortresses.<br><strong>Code is Law: </strong>Scraping is legally protected but the technical walls are so high that it&#8217;s still impossible, so only big companies can afford to pay to scrape anything worthwhile while startups, researchers and journalists can&#8217;t. <br><br>All of these are extreme futures&#8212;that&#8217;s the nature of the exercise&#8212;but in articulating them, including their pros and cons and likelihood, we were able to weigh the complexities of the issue and the tough tradeoffs involved in a productive way. I <a href="https://www.linkedin.com/feed/update/urn:li:activity:7445138636150566912/?originTrackingId=cmfBarv0NfyE7t0h5GLkjg%3D%3D">posted</a> the whole exercise worksheet on LinkedIn for others to use as a model because it&#8217;s a dead simple way to surface a wide range of predictions and perspectives around whatever issue it&#8217;s applied to. Check it out, and happy futuring!</p><h2><strong>Steven Soderbergh Volunteers To Get a Swirlie from Hannah Einbinder</strong></h2><p>Highlighting how polarizing generative AI has become in the film community, Emmy-winning <em>Hacks</em> star Hannah Einbinder roasted AI video creators during a press conference for that brilliant show&#8217;s final season, a roasting that nearly every Hollywood <a href="https://variety.com/2026/tv/news/hannah-einbinder-ai-creators-losers-1236706302/">trade</a> <a href="https://deadline.com/2026/04/hannah-einbinder-slams-ai-creators-losers-not-artists-1236772807/">press</a> <a href="https://www.thewrap.com/culture-lifestyle/culture/hannah-einbinder-ai-creators-losers-hacks/">outlet</a> gleefully gave its own story. <em><a href="https://www.slashfilm.com/2138871/hacks-hannah-einbinder-harsh-words-ai/">Slashfilm</a></em> originally reported the full quote:</p><blockquote><p>The people who make this stuff are losers. They&#8217;re not artists. They&#8217;re not creative. And they&#8217;ve wanted their whole lives to be special. And they&#8217;re not special&#8230;. They&#8217;re trying to rob real creative people of our gifts. And you can&#8217;t. And even if you try, you will never be cool. You guys suck. No one likes you. Anyone who&#8217;s near you is because they crave power and access over any ethical standard. You are a loser. You will never be cool. And you probably had a rolly backpack in high school. I wanna put your head in the toilet and flush.</p></blockquote><p>Almost immediately proving her generalization wrong, legendary Academy Award-winning director and definitely-not-a-loser Steven Soderbergh noted in an interview with <em><a href="https://filmmakermagazine.com/133556-interview-steven-soderbergh-the-christophers-spring-2026/">Filmmaker</a></em> that he&#8217;ll be using &#8220;a lot of AI&#8221; in his forthcoming films.</p><p>It was the interviewer who first raised the subject, noting the &#8220;horror of AI&#8221; as perhaps &#8220;too depressing to talk about,&#8221; but Soderbergh immediately broke the narrative: &#8220;It&#8217;s worth talking about what that technology might be good at,&#8221; he replied.</p><p>Specifically, Soderbergh noted that for ten of the ninety minutes in his upcoming documentary on John Lennon and Yoko Ono, &#8220;AI has been helpful in creating thematically surreal images that occupy a dream space rather than a literal space&#8230;. But like every other piece of technology, it desperately requires very close human supervision.&#8221;</p><p>He came back to the subject when asked how he&#8217;d handle the expense of a period film about the Spanish-American war that he&#8217;s currently developing. Interviewer: &#8220;With ships and everything? For which you would use&#8230;&#8221; Soderbergh: &#8220;A lot of AI.&#8221;</p><p>After the inevitable online backlash to these comments, Soderbergh doubled down in an interview with <em><a href="https://variety.com/2026/film/news/steven-soderbergh-the-christophers-star-wars-ben-solo-movie-controversial-ai-comments-1236713201/">Variety</a></em>, where he said that the AI blowback &#8220;is mystifying to me.&#8221; He continued:</p><blockquote><p>I felt obligated to engage with it, to figure out what it is and what it can do. It turned out to be a very good tool for certain passages of the Lennon documentary where I needed surrealistic imagery that was impossible to shoot. It allowed me to solve a creative problem about how to visualize what John and Yoko are speaking about philosophically. Ten years ago, I would have needed to engage a visual effects house at an unbelievable cost to come up with this stuff. No longer. My job is to deliver a good movie, period. And this tool showed up at a moment when I needed it&#8230;. There are some people that I have absolute love and respect for that refuse to engage with it. That&#8217;s their privilege. But I&#8217;m not built that way. You show me a new tool. I want to get my hands on it and see what&#8217;s going on.</p></blockquote><p>The use case Soderbergh describes here is illustrative: as he notes, without AI he&#8217;d need an expensive effects house, which he certainly wouldn&#8217;t have paid to engage for a little documentary. So in this case, AI expanded creative possibilities without removing opportunities for craftspeople. But that certainly won&#8217;t be true in all cases, such that claiming to be &#8220;mystified&#8221; by the uproar seems a bit obtuse and, well, privileged. So a bit of advice, Steven: steer clear of any toilets while Hannah Einbinder is around!</p><p>(Update: just as we were finalizing this edition, <a href="https://www.hollywoodreporter.com/news/general-news/reese-witherspoon-ai-comments-instagram-reel-book-authors-1236566844/">Reese Witherspoon</a> and <a href="https://deadline.com/2026/04/sandra-bullock-encourages-hollywood-lean-into-ai-1236864020/">Sandra Bullock</a> also volunteered for Einbinder swirlies.)</p><h2><strong>Hollywood&#8217;s So Angry at AI It Can&#8217;t Spell Straight</strong></h2><p>Video generation startup Runway hosted its first <a href="https://summit.runwayml.com/">Runway AI Summit</a> in NYC on March 31st, highlighting advancements in the technology and how it&#8217;s being used creatively, and let&#8217;s just say it was received&#8230;skeptically by the Hollywood and tech press.</p><p><em><a href="https://www.wired.com/story/thank-you-for-generating-with-us-hollywoods-ai-acolytes-stay-on-the-hype-train/">Wired</a></em>&#8216;s mocking headline was representative: &#8220;&#8216;Thank You for Generating With Us!&#8217; Hollywood&#8217;s AI Acolytes Stay on the Hype Train.&#8221; Meanwhile, <em><a href="https://www.hollywoodreporter.com/business/digital/kathleen-kennedy-star-wars-ai-runway-think-1236553421/">The Hollywood Reporter</a></em>&#8216;s story focused almost solely on the cautionary note sounded by one star producer during the event: &#8220;Kathleen Kennedy Just Told an AI Conference She&#8217;s Not So Sure About AI,&#8221; announced the headline.</p><p>Kennedy, who is wrapping up her tenure as the head of <em>Star Wars</em> production company LucasFilm, rather sensibly highlighted the importance of preserving human taste and the serendipitous unpredictability of the creative process in the face of AI. She also reasonably pressed for more transparency around model development, the lack of which is a particular challenge for copyright holders seeking to prevent their content from being trained on without consent. Then she went on a bit of a tangent criticizing the quality of 3D-printed props compared to those made by craftspeople.</p><p><em>THR</em>&#8216;s story contained an embarrassing mistake that hopefully isn&#8217;t emblematic of their team&#8217;s actual level of tech knowledge. The article referred to ByteDance&#8217;s Seedance 2.0 model&#8212;the same model that has caused no end of controversy and flurry of cease-and-desist letters from the content industry since its preview release in February&#8212;as &#8220;ByteDance&#8217;s Seesaw.&#8221; That mistake wasn&#8217;t corrected for several days (despite this author&#8217;s repeated <a href="https://x.com/KevinBankston/status/2040037005472510116">nagging on X</a>). And the story still weirdly calls Seedance a &#8220;social application,&#8221; which, just, no.</p><p>I&#8217;m surprised that I haven&#8217;t yet seen sarcastic press coverage of <a href="https://www.hollywoodreporter.com/movies/movie-news/soulscape-2026-ai-filmmakers-cinema-lab-summit-1236547017/">SoulScape</a>, a new &#8220;Global AI Cinema Lab and Summit&#8221; that took place in San Francisco a weekend ago, but don&#8217;t worry, there are plenty of upcoming AI+film events for <em>Wired</em> and <em>THR</em> to make fun of. Next up: Runway will be hosting its <a href="https://aif.runwayml.com/">fourth annual AI film festival</a> in both NYC and LA in early June.</p><p>Taking the Einbinder comments and the chilly reception to the Runway event together, they seem representative of the problem with the Hollywood discourse on AI as diagnosed by <em><a href="https://theankler.com/p/sfs-ai-billboards-should-terrify">The Ankler</a></em>&#8216;s chief columnist Richard Rushfield:</p><blockquote><p>The conversation, such as it is, shoves everything in one basket and reacts to it all with a primal scream, lumping together theft and innovation, job loss and useful tools, corporate abuse and creative experimentation&#8230;. The working premise in a lot of Hollywood is that AI is <em>evil and must go away</em>. Anything less than total resistance is surrender. I understand the impulse.... But that <a href="https://x.com/TTomTToro/status/2034661965231198616">isn&#8217;t a strategy</a>. [italics in original; funny link added.]</p></blockquote><p>In an exercise evoking the 2x2 matrix, Rushfield wisely urged Hollywood to carefully delineate between four different categories of AI impacts&#8212;good and unstoppable, good and stoppable, bad and unstoppable, bad and stoppable&#8212;to better focus on stopping the bad things it can rather than wasting effort on the bad things it can&#8217;t, taking advantage of the good things that are coming, and avoiding blocking the stoppable benefits. Let&#8217;s see if anyone takes his advice.</p><h2><strong>Pay No Attention to the Foundation Model Behind Ben Affleck&#8217;s Curtain</strong></h2><p>One recent development at the intersection of AI and the entertainment industry somehow escaped the primal scream of the anti-AI contingent. That was Netflix&#8217;s high-profile <a href="https://about.netflix.com/en/news/why-interpositive-is-joining-netflix">acquisition</a> of Ben Affleck&#8217;s stealth AI startup, InterPositive, in a deal that ultimately could be worth up to $600 million for the actor. Despite the highly polarized debate over generative AI in Hollywood, Affleck and Netflix effectively sidestepped controversy by describing their AI tools as filmmaker-centric and implying that they were trained using only small sets of proprietary filmed data. But it looks like they got off easy, by not being fully up-front about how their technology works.</p><p>Rather than a prompt-to-video-slop engine, InterPositive&#8217;s tools are solely focused on improving real footage that&#8217;s already been shot, consistent with a human director&#8217;s vision. &#8220;For artists to apply these [AI] tools towards telling the stories we dedicate our lives to,&#8221; Affleck said in the <a href="https://about.netflix.com/en/news/why-interpositive-is-joining-netflix">press release</a>, &#8220;they need to be purpose-built to represent and protect all the qualities that make a great story&#8230;. [We] need to preserve what makes storytelling human, which is judgment&#8230;. I knew I had a responsibility to my peers and our industry, to protect the power of human creativity and the people behind it. In creating InterPositive, I sought to do just that.&#8221;</p><p>This is all consistent with Affleck&#8217;s <a href="https://variety.com/2024/film/news/ben-affleck-slams-ai-wont-replace-movies-1236213287/">prior statements</a> (before anyone knew he was secretly working on an AI startup) where he highlighted AI&#8217;s limits compared to human creativity but also cautiously described its potential creative uses for filmmakers: &#8220;What AI is going to do is going to dis-intermediate the more laborious, less creative, and more costly aspects of filmmaking, that will allow costs to be brought down, that will lower the barrier to entry, that will allow more voices to be heard, that will make it easier for the people want to make <em>Good Will Hunting</em>s, to go out and make it.&#8221;</p><p>In describing InterPositive&#8217;s AI models, Affleck and Netflix highlighted their artisanal nature (all emphases added): &#8220;I began filming a <em>proprietary dataset</em> on a controlled soundstage with all the familiarities of a full production&#8230;.The results of this foundational work were <em>deliberately smaller datasets and models</em> focused on filmmaking techniques &#8212; rather than [generating] performances &#8212; creating tools that artists can use, control and benefit from.&#8221;</p><p>As Affleck further explained to <em><a href="https://variety.com/2026/digital/news/netflix-600-million-ben-affleck-ai-film-startup-interpositive-1236685038/">Variety</a></em>, &#8220;[t]he InterPositive system builds an AI model based on <em>an existing production&#8217;s dailies</em>, then lets a filmmaker introduce that model into the postproduction process to provide the ability to do things like mix and color, relight shots, and add visual effects.&#8221;</p><p>These statements imply (but never state) that InterPositive&#8217;s models rely only on small amounts of proprietary data. That narrative never made much sense technically, since generative AI models require enormous amounts of data. For example, Lionsgate&#8217;s entire catalogue <a href="https://www.thewrap.com/lionsgate-runway-ai-deal-ip-model-concerns/">wasn&#8217;t enough</a> to build a useful custom video model in that studio&#8217;s partnership with Runway. But that didn&#8217;t stop other outlets from uncritically running with and expanding on these descriptions, describing InterPositive&#8217;s technology as being refreshingly free from unconsented use of copyrighted works in its training, in contrast to bigger, general-purpose video generation models:</p><ul><li><p>&#8220;The technology, however, <em>doesn&#8217;t&#8230;use footage without permission</em>. Additionally, the AI model is getting trained on <em>material you already own and have access to</em>.&#8221; &#8211;<em><a href="https://www.inc.com/leila-sheridan/ben-affleck-just-sold-his-stealth-ai-startup-to-netflix-for-600-million-heres-what-it-actually-does/91315812">Inc.</a></em></p></li><li><p>&#8220;InterPositive trains a custom AI model on <em>a production&#8217;s own dailies</em>, using that footage as the foundational dataset <em>rather than pulling from public internet sources</em>.&#8221; &#8211;<em><a href="https://www.cined.com/netflix-acquires-ben-afflecks-interpositive-ai-filmmaking-company/">CineD</a></em></p></li><li><p>&#8220;Models are trained exclusively on <em>first-party footage rather than third-party datasets</em>, [avoiding] intellectual property and consent risks&#8230;.&#8221; &#8211;<em><a href="https://delmorganco.com/netflix-moves-to-acquire-interpositive-in-600mm-deal/">Del Morgan &amp; Co.</a></em></p></li><li><p>&#8220;Netflix&#8217;s acquisition of InterPositive signals a deliberate pivot toward proprietary AI tools designed specifically for filmmaking <em>rather than relying on general-purpose generative AI models</em>.&#8221; &#8211;<em><a href="http://mlq.ai">MLQ.ai</a></em></p></li><li><p>&#8220;Affleck said InterPositive <em>did not provide video generation tools such as Google&#8217;s Veo3 or OpenAI&#8217;s Sora</em> &#8211; it was &#8220;not about text prompting or generating something from nothing&#8221; &#8211; but instead helped in the post-production process.&#8221; &#8211;<em><a href="https://www.theguardian.com/technology/2026/mar/06/ben-affleck-sells-ai-postproduction-startup-interpositive-to-netflix">The Guardian</a></em></p></li></ul><p>However, the actual <a href="https://patents.google.com/patent/US12438995B1/en?inventor=Benjamin+Geza+Affleck-Boldt">patent</a> for InterPositive&#8217;s technology, published by <em><a href="https://deadline.com/2026/04/netflix-ben-affleck-ai-firm-interpositive-film-production-savings-1236770381/">Deadline</a></em> a couple weeks ago, tells a different story. <em>Deadline</em>&#8216;s news hook for the story was highlighting the millions of dollars of production cost savings that the patent application promised for filmmakers, including 50% savings on visual effects. They contrasted these cost-cutting promises with other Affleck statements focused on making filmmaking &#8220;easier&#8221; and &#8220;faster&#8221; instead of focusing on cost savings, and a Netflix exec&#8217;s statement that &#8220;it&#8217;s not really about cheaper, it&#8217;s really about better.&#8221; Obviously it&#8217;s about both, so not much of a gotcha there. But there was another story buried in the patent that <em>Deadline</em> missed:</p><p>Contrary to what Affleck and Netflix have implied, InterPositive&#8217;s technology must be built on top of general-purpose generative AI models that typically have been trained on masses of copyrighted material.</p><p>In particular, the patent reveals that InterPositive&#8217;s proprietary tools and data are meant to run on top of another large video gen model being used as the foundation of the system, essentially &#8220;fine-tuning&#8221; the larger base model by transferring learning from the smaller InterPositive model trained on filmmaking-specific knowledge. As I highlighted on <a href="https://x.com/KevinBankston/status/2040099537964966112">Twitter</a> when I discovered it, the patent specifically points to OpenAI&#8217;s Sora and Google&#8217;s video model as examples of appropriate base models. From the patent:</p><blockquote><p>[InterPositive&#8217;s model is used] to train the other model, enabling it to recognize and replicate filmmaking techniques when provided with appropriate prompts&#8230;. [That] video model may be an existing large language model, such as OpenAI SORA or a Google AI model, which has been primarily designed for processing and generating video content. Prior to the transfer learning process, these models lack the capability to accurately interpret and implement cinematographic details in their outputs&#8230;. The video LLM serves as the base model that is enhanced through transfer learning to acquire the advanced filmmaking capabilities developed in [InterPositive&#8217;s] model.</p></blockquote><p>I say that InterPositive&#8217;s tools are &#8220;meant to run&#8221; on top of bigger models rather than they &#8220;do run&#8221; because at this point it&#8217;s still not clear that they even exist in a workable form right now. <em><a href="https://www.thewrap.com/industry-news/deals-ma/netflix-acquires-ben-afflecks-ai-production-startup/">The Wrap</a></em> reports that the model is still in development, and as one Netflix executive put it, &#8220;[i]t&#8217;s not like it&#8217;s a complete and ready to go tool.&#8221; But wherever it is in the R&amp;D process, what InterPositive is developing is a method for fine-tuning a larger foundation model with specialized data for specific use in film postproduction, rather than building its own independent model trained only on proprietary data.</p><p>To be clear: I&#8217;m not judging InterPositive for choosing to build on top of state-of-the-art foundation models from major US AI labs in order to develop creator-centric, film-focused tools. And I&#8217;m a fan of Affleck, a wildly talented, Academy Award-winning actor-writer-director-producer, as much as any other Gen Xer film nerd who religiously watched every episode of <em>Project Greenlight</em> and always cries at the end of <em>Armageddon</em> can be.</p><p>What I am judging is his and Netflix&#8217;s obscuring that fact rather than standing by it and explaining their choices. Having true creators champion reasonable, responsible, limited uses of this technology in the creative process means something in this conversation; subtle attempts to shy away from doing that in order to avoid criticism mean something, too. Perhaps Affleck should take a cue from Soderbergh and fellow AI-forward director <a href="https://deadline.com/2026/02/darren-aronofsky-auteur-ai-youtube-backlash-1236709982/">Darren Aronofsky</a>, both of whom have braved the backlash to defend their creative use of these tools, rather than avoid the hard conversation.</p><p>In furtherance of that hard conversation, the key question remains and should be answered directly by Affleck and Netflix: <em>on top of what foundation model, exactly, is Netflix going to be building InterPositive&#8217;s tools?</em></p><h2><strong>Director Robert Rodriguez on AI as Creative Multiplier</strong></h2><p>Speaking of true creators, last month I had the unique pleasure of visiting the film industry equivalent of Willy Wonka&#8217;s Chocolate Factory: Robert Rodriguez&#8217;s Troublemaker Studios in Austin, Texas. I was there as an investor in his new development company <a href="https://brassknucklefilms.com/">Brass Knuckle Films</a>, cofounded with producer Alexis Garcia. Like Eli Roth&#8217;s <a href="https://horrorsectionstudios.com/">The Horror Section</a>, BKF is funded through blockchain-based securities issued through <a href="http://republic.com">Republic.com</a>, such that investors will receive a share of any profits from the ventures. This new convergence of technology-enabled securities and film financing is itself an interesting development, but I was even more fascinated by the conversations at Troublemaker.</p><p>Unsurprisingly, it was absolutely delightful to meet Robert and tour the studio: hey look, it&#8217;s the biggest standing set in the US, from <em>Alita: Battle Angel</em>! Cool, there&#8217;s the car from Quentin Tarantino&#8217;s <em>Death Proof</em>! Oooh, that&#8217;s the matte painting from the end of <em>From Dusk Till Dawn</em>! Hearing Robert share his creative philosophy and preview the studio&#8217;s latest projects was just as inspiring as you would imagine.</p><p>Most interesting was hearing Robert talk about the future of AI in film production, where he voiced his hope to become an innovator not just in how to creatively use the tools but also in defining industry guardrails around their ethical use, which was music to this AI governance and copyright nerd&#8217;s ears.</p><p>To illustrate his hope that AI will multiply rather than supplant human creative capacity, he showed how he&#8217;d taken a 2D image of a cartoon monster that he designed and then experimented with tools from Luma Labs (which was a cosponsor of Brass Knuckle&#8217;s SXSW party earlier in March) to create a 3D model of the same character in a style that matched his creative intent. Echoing Soderbergh&#8217;s comments, Rodriguez highlighted how without AI he would have needed to send the 2D image to a designer to build a model that likely wouldn&#8217;t match his imagination and would need to be sent back for additional iteration, and at great cost, while the AI allowed him to quickly generate exactly what was in his head.</p><p>For a DIY creator who likes to do as much as possible on his productions both for creative and budget reasons&#8212;he shoots, he edits, he scores!&#8212;it&#8217;s not surprising to hear Rodriguez talking this way. He has always been a uniquely tech-forward and cost-conscious filmmaker. And to be clear, he didn&#8217;t make any specific representations about whether, when, or how he&#8217;ll be introducing generative AI into his development or production pipeline; the example he shared was just him fooling around with the tools. Even so, it&#8217;s good to have another veteran director willing to at least begin to engage on the question of where we do want to use AI in film, and where we don&#8217;t.</p><p>All that and Tex-Mex too! Truly, I couldn&#8217;t have asked for a better Austin weekend.</p><h2><strong>Virtual Studios: What&#8217;s the Difference Between </strong><em><strong>Sin City</strong></em><strong> and Generated Cities?</strong></h2><p>The new AI-driven virtual studio&#8212;the gen-AI iteration of the green-screen studios used by Robert Rodriguez on his <em>Sin City</em> films or Lucas on his <em>Star Wars</em> prequels&#8212;isn&#8217;t coming soon. Rather, it&#8217;s already here, with news this week of both a major feature film and a major streaming series soon to be released.</p><p>First up is <a href="https://www.thewrap.com/creative-content/movies/ai-movie-bitcoin-killing-satoshi-gal-gadot-casey-affleck-doug-liman/">news</a> of director Doug Liman (<em>The Bourne Identity</em>, <em>Mr. &amp; Mrs. Smith</em>) completing principal photography on <em>Bitcoin: Killing Satoshi</em>, his $70 million (?!) generative AI feature about the mysterious inventor of the cryptocurrency. Starring Gal Gadot, Pete Davidson, Isla Fisher, and Casey Affleck (AI runs in the family!), this &#8220;globe-trotting thriller&#8221; with over 200 locations purportedly would have cost $300 million if shot IRL. Instead, though, it was shot in a small gray warehouse with generated backgrounds to be added later.</p><p>Obviously that script would never have been produced as written and locations would&#8217;ve been pared down on a traditional shoot, so that <em>Avatar</em>-level number is a bit ridiculous. And the claim that this will be the first 100% generated film is also a bit misleading even if technically true. Presumably, every pixel will indeed be generated&#8212;including reproduction of the actors performances, in a video-to-video (as opposed to text-to-video) generative pipeline&#8212;rather than recorded performances simply being composited on top of generated backgrounds. Their process likely looks more like that used by this new and technically impressive <a href="https://x.com/KevinBankston/status/2044618748682719695">AI-generated Spanish horror short film</a> completed with consumer-grade tools (which raises the question of how this thing still cost $70 million). But saying &#8220;fully AI-generated&#8221; evokes completely generated performers as well, which this definitely isn&#8217;t.</p><p>Next up is the similar <a href="https://lumalabs.ai/news/luma-innovative-dreams">announcement</a> from Luma Labs and production company Wonder Project, the producers of the Amazon streaming hit <em>House of David</em>, that they are teaming up to launch new AI studio called Innovative Dreams. That new studio is making Wonder&#8217;s next biblical series <em>The Old Stories: Moses</em> (with Ben Kingsley as the prophet) using what they are calling &#8220;hybrid filmmaking,&#8221; which appears to be the same sort of video-to-video performance capture and generative pipeline as the other projects mentioned above.</p><p>Finally, in a reversal of these stories about real actors and generated backgrounds, there was this past week&#8217;s news about a generated actor inserted into a traditional shoot: the digital &#8220;resurrection&#8221; by AI of deceased actor Val Kilmer in the <a href="https://variety.com/2026/film/news/val-kilmer-ai-as-deep-as-the-grave-trailer-1236722342/">trailer</a> for the upcoming feature <em>As Deep as the Grave</em>. Although legally blessed by the actor&#8217;s family, this development raises the uncomfortable prospect that the digital likenesses of stars, like music libraries, are becoming something akin to tradable <a href="https://theankler.com/p/val-kilmer-and-immortal-ip-stars">financial instruments</a>.</p><p>This isn&#8217;t just a story about entertainers who have passed away, either: as Taylor Lorenz reported this past week in <em><a href="https://www.vanityfair.com/news/story/influencers-ai-clones">Vanity Fair</a></em>, online influencers are increasingly producing content using generated &#8220;AI clones.&#8221; As one online commentator <a href="https://x.com/TaylorLorenz/status/2044596152352575839">noted</a>, &#8220;there will be very little reason for a Mr. Beast to actually show up to film a super bowl commercial in a few years [since] his identity will be in a .zip file controlled by [his manager].&#8221;</p><p>Again, all of these stories received an enormous amount of backlash online, but regarding <em>Bitcoin</em> and <em>Moses</em>, I&#8217;m left wondering: how is this so different from what Lucas and Rodriguez were already doing in the early 2000s? Does it really matter what kind of software is being used to fill in the backgrounds? Is the concern that these new methods reasonably feel more likely to meaningfully replace location shooting and the jobs that go with them, compared to those previous experiments that didn&#8217;t radically change the production ecosystem? (That&#8217;s a questionable assertion when you look at how omnipresent green screens are in modern big-budget productions.)</p><p>The concern makes some sense, but is it necessarily a bad thing for a domestic film industry that has lost much of its location shooting to foreign shoots anyway? It&#8217;s not clear to me that being able to shoot many more low-budget features in a studio, perhaps even in LA again, will employ fewer American cast and crew than producing much fewer big-budget on-location features internationally with the same amount of money. But we&#8217;ll definitely need to get that $70 million number down first, and get closer to the &#8220;make fifty movies with $100 million instead of one movie&#8221; future that Runway&#8217;s CEO is <a href="https://techcrunch.com/2026/04/16/runway-ceo-says-ai-could-help-hollywood-make-50-films-instead-of-one-100m-blockbuster/">promising</a>. And even then, the jobs saved in LA will still cost jobs elsewhere.</p><h2><strong>So Long Sora, Welcome Back Seedance 2.0, Hello to Netflix&#8217;s VOID</strong></h2><p>All this talk of AI-based production raises the question: what models will Hollywood be using to drive their virtual studios? This past month answered that question in regard to at least one model: it definitely <a href="https://x.com/KevinBankston/status/2036638117562585347">won&#8217;t be OpenAI&#8217;s Sora</a>!</p><p>When OpenAI <a href="https://x.com/soraofficialapp/status/2036546752535470382">announced</a> on March 24th that it was <a href="https://help.openai.com/en/articles/20001152-what-to-know-about-the-sora-discontinuation">shutting down</a> Sora entirely&#8212;app dead by April 26th, API dead by September 24th, Disney&#8217;s $1 billion licensing deal <a href="https://www.reuters.com/technology/openai-set-discontinue-sora-video-platform-app-wsj-reports-2026-03-24/">collapsing</a> within an hour of the news&#8212;it wasn&#8217;t just a bad day for the &#8220;Hollywood&#8217;s cooked&#8221; crowd on AI Twitter. It was also a load-bearing wall getting yanked out of any Hollywood AI pipeline that had bet on that model as a foundation (potentially including InterPositive, based on their patent). If your post-production toolchain depends on a model whose existence can be terminated overnight by a vendor whose incentives don&#8217;t align with yours, you don&#8217;t have a pipeline, you have a problem.</p><p>Meanwhile, as OpenAI&#8217;s Sora is retiring, ByteDance&#8217;s controversial Seedance 2.0 model is returning. The aggressively copyright-infringing preview of that model was pulled last month after a flurry of cease-and-desist letters from Hollywood. But it finally had its official release in the United States on April 9th, first via ByteDance&#8217;s video editing platform <a href="https://www.capcut.com/newsroom/dreamina-seedance-2">CapCut</a> and then via other third-party model platforms, after a rolling international release.</p><p>However, the version that landed here bears only a passing resemblance to the one that generated the now infamous <a href="https://www.dailymotion.com/video/x9ztevg">Tom Cruise vs. Brad Pitt fight clip</a>. The Chinese-domestic Jianying version offers photorealistic faces, multi-shot storytelling, and almost no guardrails. The Dreamina Seedance 2.0 that paid CapCut Pro subscribers in the US can now access has been comprehensively defanged: real-face uploads are blocked, intellectual property keywords and visual matches are filtered, every output is watermarked, clips are capped at fifteen seconds, and the content moderation is aggressive enough that some creators are calling it &#8220;<a href="https://x.com/zabiisuto/status/2025922391508197875">nerfed</a>.&#8221; So now that Sora is dead, this potential back-up vendor for Hollywood&#8217;s AI pipelines is delivering a fraction of the capability that made it interesting in the first place.</p><p>Which makes the third story from the same news cycle perhaps the most interesting. On April 3rd, Netflix quietly released <a href="https://huggingface.co/netflix/void-model">VOID</a>&#8212;Video Object and Interaction Deletion&#8212;on Hugging Face under an Apache 2.0 license, its first-ever open-source AI model. VOID isn&#8217;t a Sora replacement; it&#8217;s a specialized fine-tuned model built on top of CogVideoX, the open source version of Chinese lab Zhipu AI&#8217;s proprietary Qingying video generation model. VOID narrowly focuses on physics-aware object removal, such that if you take someone or something out of a shot, VOID will also remove any physical impacts prompted by the now-absent person or item.</p><p>But the context is what matters: the same company that just spent up to $600 million acquiring the proprietary InterPositive workflow is also using and releasing its own open source video models. Those moves look like the beginnings of an overall AI pipeline strategy: avoiding reliance on outside proprietary models and instead developing internal proprietary models and using flexible open source models that won&#8217;t change or disappear at a moment&#8217;s notice.</p><p>The lesson for every studio and indie shop drafting its 2026 AI strategy (especially Disney who just got royally burned by OpenAI): better to control your own tech stack wherever possible, rather than rely on outside models that may not survive the brutal competition between labs, and where even if they do survive, the version you&#8217;re allowed to use may not be the version you wanted.</p><h2><strong>Writers vs. AI vs. Writers</strong></h2><p>Turning away from Hollywood and toward the publishing world, the last three weeks of March 2026 will be remembered as the moment when AI-generated writing stopped being a hypothetical across every institution that publishes words for a living. Five stories broke in tight succession, all variations on a single theme: who actually wrote that, and how much does the answer matter?</p><p>Let&#8217;s start with the AI resistance. On March 19th, Hachette canceled the US release of Mia Ballard&#8217;s horror novel <em>Shy Girl</em> through its Orbit imprint and discontinued the existing UK edition, after a <em>New York Times</em> <a href="https://www.nytimes.com/2026/03/19/books/shy-girl-book-ai.html">investigation</a> (and a Pangram analysis flagging ~78% of the text as AI-generated) forced their hand. This is the first known instance of a Big Five publisher walking back a book discovered to be written by AI. That discovery was driven by a user on r/horrorlit who catalogued the linguistic tells (the word &#8220;sharp&#8221; appearing 186 times, lists of three, em-dash overuse), then escalated through a 1.4-million-view <a href="https://www.youtube.com/watch?v=GbeKTa5xhZo">YouTube essay</a> titled &#8220;i&#8217;m pretty sure this book is ai slop.&#8221;</p><p>The next day, the <em>New York Times</em> <a href="https://www.thewrap.com/media-platforms/journalism/new-york-times-cuts-ties-with-writer-ai/">severed ties</a> with freelance critic Alex Preston after a reader noticed that his January book review had lifted phrases and full paragraphs from a <em>Guardian</em> critic&#8217;s <a href="https://www.theguardian.com/books/2025/aug/21/watching-over-her-by-jean-baptiste-andrea-review-a-love-song-to-italy">review</a> of the same novel. Preston confessed he&#8217;d used an AI tool to draft the review and &#8220;failed to identify and remove overlapping language from another review that the AI dropped in&#8221;&#8212;a new failure mode adjacent to plagiarism, where the model is the laundering machine.</p><p>Then on March 26th came a one-two punch from the &#8220;AI can be helpful for writers and publishers, actually&#8221; side of the debate. The <em>Wall Street Journal</em> <a href="https://www.wsj.com/business/media/an-ai-upheaval-is-coming-for-media-this-journalist-is-already-all-in-3511d951">profiled</a> a <em>Fortune</em> business editor who has cranked out 600+ articles since July 2025 using Perplexity and NotebookLM, accounting for ~20% of <em>Fortune</em>&#8216;s web traffic in H2 2025. <em>Fortune</em>&#8216;s EIC told the <em>Journal</em> &#8220;more than 50% is Nick&#8221;&#8212;a quote that is perhaps less reassuring than it was probably intended to be.</p><p><em>Wired</em> dropped a same-day companion <a href="https://www.wired.com/story/tech-reporters-using-ai-write-edit-stories/">profile</a> of independent tech journalist Alex Heath, who uses Claude Cowork (with a custom skill called &#8220;10 Commandments for Writing Like Alex Heath&#8221;) to draft his Substack pieces end-to-end. Journalists across the internet variously attacked and celebrated these practices, fanning the flames of a fire that started a few weeks before when the editor of the <em><a href="https://www.cleveland.com/news/2026/02/three-waves-of-reaction-to-our-ai-experiment-letter-from-the-editor.html">Cleveland Plain Dealer</a></em> promoted that newsroom&#8217;s use of AI to help source and write stories.</p><p>Sitting underneath all four new stories was <em>NYT</em>&#8216;s March 9th <a href="https://www.nytimes.com/interactive/2026/03/09/business/ai-writing-quiz.html">interactive quiz</a>&#8212;&#8220;Who&#8217;s a Better Writer: A.I. or Humans?&#8221;&#8212;a kind of Voight-Kampff test for prose, in which 86,000 readers blindly judged five paired passages and 54% picked the AI-generated ones, with one pairing splitting 67-33 in the machine&#8217;s favor.</p><p>Five stories, three weeks, one underlying question: is using AI to write a craft choice or a category violation? Some writers say one thing, some say the other, and 86,000 quiz-takers suggest that readers may not care which.</p><p>Full disclosure for CONVERGER readers: the entirety of this newsletter was originally written by me without AI assistance except for this feature and the previous one about Sora, where I experimented with a different workflow. For these two features, I solicited first drafts from Claude (inputting an outline with the key facts, key sources, and my rough take, then providing the previous features as an example of my style), then I heavily edited them (&#8221;more than 50% is Kevin&#8221;).</p><p>Did you see a difference as you were reading? The <a href="https://www.pangram.com/">Pangram</a> AI text detector didn&#8217;t: it concluded with a medium level of confidence that those sections were 100% human-written. Then again, the same tool <a href="https://x.com/onbeinganangel/status/2036892974508843512">predicted</a> that a paragraph from Mary Shelley&#8217;s Frankenstein was 100% AI-written. So I&#8217;d love to hear your feedback on whether you noticed a shift in style.</p><h2><strong>A Contract That Helps Protect Comic Artists Against AI Training on Their Work</strong></h2><p>Comic book artists are rightfully concerned about AI being used to copy their styles or their designs and put them out of work. That&#8217;s a problem for someone like me, who works in AI but also is a writer currently developing a number of comic book projects and looking for professional artists to work with. Therefore, to reassure the artists with whom I&#8217;ve directly contracted to develop my comic ideas, I wrote the following anti-AI-training clause for our agreements:</p><blockquote><p>Writer shall not use, nor license or authorize third parties to use, the Services [i.e., the artwork] in any manner for purposes of training generative artificial intelligence (AI) models to generate text, images, video or audio, including models to generate images or video reproducing the artwork or art style of the Artist, without prior written consent of the Artist. Nothing in this clause shall prohibit the use of generative AI models for routine internal production purposes including format conversion, upscaling, accessibility re-flows, translation, palette adjustment, or comparable non-creative functions.</p><p>[Another clause similarly prohibits the Artist from training on the commissioned artwork or my IP without my consent.]</p><p>&#8220;Generative AI models&#8221; include large language models, transformer models, diffusion models, and any other substantially similar types of machine learning models now or in the future that generate text, image, video or audio outputs based on model parameters derived from training on large datasets, regardless of whether those models are offered publicly, shared privately, or used internally, and regardless of whether those models are offered with open source licenses, closed source licenses, or are wholly proprietary. &#8220;Training&#8221; includes both the pre-training process that generates a model&#8217;s parameters and any post-training fine-tuning of a model&#8217;s parameters.</p><p>Artist and Writer each acknowledge that the other is not responsible for the unauthorized actions and conduct of third parties who attempt to use the Services or IP for generative AI training.</p></blockquote><p>Remember: I&#8217;m not your lawyer! Please talk to one before sticking this language in your contracts. And if you have suggestions on how to improve this, let me know.</p><div><hr></div><h1><strong>FRAGMENTS</strong></h1><h2><strong>WGA Deal on AI: Is That All There Is?</strong></h2><p>Beyond the narrow set of copyright and rights-of-publicity harms that litigation might address for certain actors and creators, organized labor is easily the most powerful lever for protecting entertainment industry workers from seismic displacement by AI. Which makes it all the more disappointing that the draft deal between the Writers Guild of America and the studios contains basically nothing new on AI, despite the WGA&#8217;s <a href="https://variety.com/2026/film/news/wga-ai-training-amptp-talks-1236684012/">starting demand</a> of payment for training on guild members&#8217; scripts. As <em><a href="https://variety.com/2026/film/news/wga-deal-321-million-health-fund-residuals-ai-1236711731/">Variety</a></em> first reported, &#8220;the deal largely preserves the status quo&#8221; on AI, with the studios only agreeing &#8220;to continue to hold meetings with the WGA, and to notify the guild if it licenses writers&#8217; work for AI training.&#8221; Hopefully the other guilds, including the Screen Actors Guild, will fare better in their negotiations.</p><h2><strong>News of the Supreme Court&#8217;s Ruling on AI and Copyright Has Been Greatly Exaggerated</strong></h2><p>On March 2 the Supreme Court <a href="https://www.scotusblog.com/cases/case-files/thaler-v-perlmutter/">declined to review</a> the <a href="https://media.cadc.uscourts.gov/opinions/docs/2025/03/23-5233.pdf">DC Circuit&#8217;s decision</a> in the case of <em>Thaler v. Perlmutter</em> that copyright does not protect works that are generated by AI without any human input; i.e., there must be a human author. This non-decision by the Supremes was immediately and widely touted by anti-AI voices online for the broad proposition that the Supreme Court had ruled that AI-generated content isn&#8217;t copyrightable. However, the facts of this case were really narrow: the AI at issue wasn&#8217;t modern gen AI but rather an AI art project whereby an algorithm trained on a wide swath of human works of art would self-generate new works with no human instruction or involvement at all; it was literally press-here-to-generate.</p><p>The much harder question of how much human guidance is required to create a copyright interest in modern AI-generated content&#8212;how detailed or iterative the prompts, how much human-created material uploaded as reference or for generative modification, how much human editing and arrangement of the outputs, etc.&#8212;still has yet to be answered by the courts. So, what&#8217;s the safe bet right now when working with AI? <a href="https://www.bakerdonelson.com/supreme-court-denies-certiorari-in-thaler-v-perlmutter-ai-cannot-be-an-author-under-the-copyright-act">Lawyers</a> suggest relying as much as possible on your own creative output (hopefully you were already doing that!), and documenting exactly what you personally authored or modified.</p><h2><strong>Macro-Growth in the Micro-Drama Content Pipeline</strong></h2><p>Micro-dramas are soapy, cliffhanger-heavy, short-form vertical TV series with sensational hooks (<em>The Double Life of My Billionaire Husband</em>; <em>Tricked Into Having My Ex-Husband&#8217;s Baby</em>; <em>Return of the Abandoned Heiress</em>) that are optimized for compulsive mobile viewing. Already a mature multi-billion business in Asia, the US market has finally started to <a href="https://variety.com/2025/tv/news/global-microdrama-boom-1236560947/">catch up</a> in the past couple years, including an <a href="https://www.thewrap.com/media-platforms/tv/behind-the-scenes-microdrama-set-visit/">explosion</a> of low-budget microdrama production in Los Angeles and a recent forty-microdrama production partnership between <a href="https://www.thewrap.com/media-platforms/tv/fox-entertainment-dhar-mann-vertical-dramas-deal/">YouTube star Dhar Mann</a> and Fox Entertainment. Now, in just the past few weeks, a number of major content producers have made deals to adapt or produce original work for this exploding medium: <a href="https://www.publishersweekly.com/pw/by-topic/industry-news/publisher-news/article/100065-harpercollins-partners-with-ai-powered-animation-house-toonstar.html">HarperCollins</a> partnered with AI-driven animation studio Toonstar to produce a slate of animated microdramas, starting with popular middle grade series <em>Friendship List</em>, while its romance imprint Harlequin signed a similar deal with another animation tech studio, Dashverse; <a href="https://www.thewrap.com/creative-content/tv-shows/issa-rae-microdrama-screen-time-tiktok-partnership-pinedrama/">Issa Rae&#8217;s Hoorae Media</a> inked a deal to bring microdramas to TikTok, starting with the horror series <em>Screen Time</em>; and finally the <a href="https://hollywoodreporter.com/business/business-news/the-national-enquirer-produce-microdramas-1236558128/">National Enquirer</a> is licensing its archives for a microdrama slate with verticals app GammaTime. I can&#8217;t wait for the inevitable Batboy microdrama!</p><h2><strong>Meanwhile, In the YouTube-to-Theaters Pipeline: </strong><em><strong>Backrooms</strong></em><strong> is Coming</strong></h2><p>Just a few months after YouTube star Markiplier (Mark Fischbach) had modest but definite success with his self-financed and self-produced sci-fi feature <em><a href="https://www.bbc.com/news/articles/c2lg5r414wpo">Iron Lung</a></em>, the trailer for the next YouTube-to-feature title has hit. From A24, it&#8217;s called <em><a href="https://variety.com/2026/film/news/backrooms-trailer-horror-youtube-a24-1236670649/">Backrooms</a></em>, based on a YouTube series of horror shorts inspired by a creepypasta meme that originated on 4chan, about being trapped in an eerie dimension of endless empty office hallways. The director of the shorts is making his feature debut, and the trailer appropriately vibes creepy as hell.</p><h2><strong>Old Music Beats New Music Beats AI Music</strong></h2><p>Despite concerns about endless new AI-generated music tracks swamping streaming sites, the latest data shows that old music is absolutely dominating over all new music: data from <a href="https://x.com/anishmoonka/status/2040208223421006167">Luminate</a> shows that new music (less than 18 months old) accounted for only 35.8% of what Americans listened to in 2024, compared to 73.3% in 2014, while according to <a href="https://hmc.chartmetric.com/email/af245904-ae71-4883-8595-83a811445bc6/">ChartMetric</a>, only 3 of the top 10 songs in 2025 were released in 2025. Meanwhile, streaming demand for original non-AI music&#8212;at least according to Universal Music Group&#8212;is not meaningfully being affected by the rise of Suno and other AI music apps. That&#8217;s what a senior UMG VP <a href="https://natlawreview.com/article/universal-music-group-may-have-admitted-dilution-theory-copyright-infringement-case">said</a> during a recent earnings call, which was probably an unpleasant surprise for UMG&#8217;s lawyers: they are suing Suno arguing that music from Suno&#8217;s AI models, trained on UMG artists&#8217; music, is diluting the streaming market for the record company&#8217;s songs. Interesting litigation strategy!</p><h2><strong>New Script-Reviewing AI Really Wants Brett Ratner to Direct Your Movie</strong></h2><p><em>The Wrap</em> tried out <a href="https://www.thewrap.com/industry-news/tech/quilty-ai-tool-script-test-does-it-work-analysis/">Quilty</a>, a new AI tool promising to analyze and provide helpful creative and business feedback on movie scripts, by feeding it the screenplays for already-produced movies <em>Sinners</em>, <em>Barbie</em>, <em>Christy</em>, and <em>Die Hard</em>. Quilty predicted that the hits would flop and the flops would hit. More bizarrely, the AI script reviewer recommended alleged sex harasser and proven hack Brett Ratner (<em>Rush Hour</em>, <em>Rush Hour 2</em>, <em>Rush Hour 3</em>), whose most recent film was the <em>Melania</em> documentary, as a top director pick for three of the four submitted scripts&#8212;including <em>Barbie</em>?! What, do the Ellisons own this company, too?</p><h2><strong>Is Fan-Created Content Supplanting Canonical Content?</strong></h2><p>According to a December study from Ogilvy Consulting cited in this excellent <em>Ankler</em> story discussing the <a href="https://theankler.com/p/mario-and-the-new-fandom-flywheel?r=13l5g&amp;utm_medium=ios&amp;triedRedirect=true&amp;_src_ref=t.co">role of fandom</a> and using the <em>The Super Mario Galaxy Movie</em> as a hook, &#8220;Two-thirds of Gen Zs spend more time with fan-created content than with the official titles.&#8221; Makes me wonder: what percentage of their consumption is non-fan, original, nonprofessional content, or even self-created content? And how quickly will those numbers grow in the next few years?</p><h2><strong>The Future&#8217;s So Bright I Have To Enter This Contest</strong></h2><p>Short story contests inviting visions of the future of a particular issue or technology are such a common feature of <a href="https://www.sciencedirect.com/science/article/pii/S0016328726000121?ref=pdf_download&amp;fr=RR-2&amp;rr=9eb2b5d1cadd3929">the sci-fi and futures toolbox</a> that I programmed <a href="https://www.youtube.com/watch?v=3QRTm7zUlGo">a panel on the topic</a> a few years ago. Yet this new <a href="https://protopianprize.com/">Protopian Prize</a> contest in particular caught my eye, though not because of the solicited topics (hopeful futures about AI and governance).</p><p>I tend to think &#8220;write more optimistic futures!&#8221; projects are based on a fallacious take that positive stories help lead to better futures than dystopias, since dystopias have arguably been a much more positive influence by providing <a href="https://www.davidbrin.com/nonfiction/tomorrowsworld.html">self-preventing prophecies</a> that help us guard against their negative futures. For example, <em>1984</em> is probably the sci-fi tale that has had the broadest positive impact on technology policy debates, including in my own experience working on privacy and surveillance policy. But the judging committee of this new prize is full of absolutely extraordinary writers and thinkers who are very aware of the value of dark future visions as well as bright, including Annalee Newitz (<em>Automatic Noodle</em>), Hannu Rajaniemi (<em>The Quantum Thief</em>), and Ruthanna Emrys Gordon (<em>A Half-Built Garden</em>, easily my favorite sci-fi novel of 2022). So, I&#8217;m very excited to see what they pick. Submissions open May 1st and close July 31st, so start futuring now!</p><h2><strong>Big Brother, Generating Slop</strong></h2><p>Speaking of <em>1984</em>, it turns out Orwell predicted <a href="https://www.uts.edu.au/news/2025/06/slopaganda-a-new-word-for-ai-spin">slopaganda</a> (AI slop for political purposes, like the <a href="https://www.theguardian.com/commentisfree/2026/apr/08/lego-videos-iran-trump-ai-video-meme-propaganda-movie-animation">mad Lego rap videos</a> coming out of Iran right now) in his classic novel. A poster on Twitter <a href="https://x.com/SamBuntz/status/2039791005050810628">highlighted</a> this passage:</p><blockquote><p>There was a whole chain of separate departments dealing with proletarian literature, music, drama, and entertainment generally. Here were produced rubbishy newspapers containing almost nothing except sport, crime, and astrology, sensational five-cent novelettes, films oozing with sex, and sentimental songs which were composed entirely by mechanical means on a special kind of kaleidoscope known as a versificator.</p></blockquote><p>Oh what a brave new world! (Wait, that&#8217;s Huxley&#8230;I mean, Shakespeare&#8230;)</p><h2><strong>Webtoon Translations, Digital Comics, Tiny Onions</strong></h2><p>Pulling together a few different comics-related items: </p><p>First up: Webtoon, a site for web-native vertical comics that has already seen <a href="https://genericpuff.tumblr.com/post/731546502706659328/on-this-weeks-episode-of-webtoon-controversies">a</a> <a href="https://www.animenewsnetwork.com/interest/2023-05-26/korean-webtoon-accused-of-using-ai-images-tracing-mushoku-tensei-anime/.198458">few</a> <a href="https://koreajoongangdaily.joins.com/2023/06/10/business/tech/Korea-webtoon-AI/20230610060020578.html">controversies</a> over allegedly AI-generated comics art, <a href="https://www.techbuzz.ai/articles/webtoon-brings-ai-translation-to-independent-comic-creators">launched</a> its opt-in AI-driven translations tool with only a <a href="https://www.webtoonish.com/p/the-webtoon-industry-killed-the-translation">tiny</a> bit of anti-AI pushback in response. </p><p>Second: digital comic book stores&#8212;the folks who sell digital versions of the paper comics sold in IRL comic shops through the comic book &#8220;direct market&#8221;&#8212;is getting crowded and confusing as a range of old and new services try to take market share from Amazon&#8217;s Comixology. Thankfully <em>The Beat</em> wrote this <a href="https://www.comicsbeat.com/a-overview-of-the-current-digital-comics-landscape/">handy guide</a> to help make sense of the current state of the digital comics landscape.</p><p>Finally: I&#8217;ll be writing a lot more someday about Tiny Onion, the comics production studio founded by star writer James Tynion IV. The young company has helped enable and solidify Tynion&#8217;s rise through the ranks of independent comics creators, with horror hits like <em>Something is Killing the Children</em> that have multiple comic spinoffs and several TV and film adaptations in the pipeline. I&#8217;d kill for a comic-nerd <em>Harvard Business Review</em> case study of Tiny Onion&#8217;s dominating entrance onto the media scene, but until then here is an <a href="https://sktchd.com/interview/eric-harburn-tiny-onion-interview/#memberful_overlay">insightful interview</a> with Eric Harburn. Harburn is Tiny Onion&#8217;s Editor-in-Chief&#8212;and now, in consideration of the company&#8217;s broad multimedia ambitions, Director of Narrative&#8212;and he talks about how Tynion and the company go about building narrative &#8220;engines&#8221; that can power multiple cross-media properties. If you&#8217;re into comics you should consider subscribing to <em><a href="https://sktchd.com/">SKTCHD</a></em>, an excellent comics news site, to read the whole thing.</p><div><hr></div><p>And that&#8217;s what&#8217;s converging this week! See you next time.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://converger.kevinbankston.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Kevin Bankston's CONVERGER! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>