<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Kamile Lukosiute]]></title><description><![CDATA[Kamile Lukosiute]]></description><link>https://blog.kamilelukosiute.com</link><image><url>https://substackcdn.com/image/fetch/$s_!ivaB!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F10501e6d-5bc0-47cc-a99c-6d176c7067ab_1280x1280.png</url><title>Kamile Lukosiute</title><link>https://blog.kamilelukosiute.com</link></image><generator>Substack</generator><lastBuildDate>Tue, 14 Apr 2026 17:31:20 GMT</lastBuildDate><atom:link href="https://blog.kamilelukosiute.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Kamile Lukosiute]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[kamilelukosiute@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[kamilelukosiute@substack.com]]></itunes:email><itunes:name><![CDATA[Kamile Lukosiute]]></itunes:name></itunes:owner><itunes:author><![CDATA[Kamile Lukosiute]]></itunes:author><googleplay:owner><![CDATA[kamilelukosiute@substack.com]]></googleplay:owner><googleplay:email><![CDATA[kamilelukosiute@substack.com]]></googleplay:email><googleplay:author><![CDATA[Kamile Lukosiute]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[American lab visibility into AI cyber misuse expires in 3–6 months]]></title><description><![CDATA[The time to invest in societal mitigations is now.]]></description><link>https://blog.kamilelukosiute.com/p/american-lab-visibility-into-ai-cyber</link><guid isPermaLink="false">https://blog.kamilelukosiute.com/p/american-lab-visibility-into-ai-cyber</guid><dc:creator><![CDATA[Kamile Lukosiute]]></dc:creator><pubDate>Fri, 27 Feb 2026 17:23:34 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/216ee239-ec75-4cf5-89a9-dfddbd169943_1862x1026.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>For cyber misuse, the frontier safety paradigm &#8211; where labs are the chokepoint &#8211; has a shelf life, and it&#8217;s about 3&#8211;6 months. Open-weight models are approaching the threshold where they&#8217;re useful for offensive cyber operations. Once that threshold is crossed, attackers will prefer using open-source models, since even a slightly worse model is worth it if it means not handing evidence of criminal activity to a US company.</p><p>This week, Anthropic <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">accused</a> Chinese AI labs Moonshot, DeepSeek, and Minimax of large-scale distillation attacks against Claude. Preventing distillation might be the highest-leverage intervention for slowing down the proliferation of offensive cyber capabilities right now. But even if we figure out how to do that, we should be building defences for a world where open-source models are catching up.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aOKM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aOKM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 424w, https://substackcdn.com/image/fetch/$s_!aOKM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 848w, https://substackcdn.com/image/fetch/$s_!aOKM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!aOKM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aOKM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg" width="498" height="493.8956043956044" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1444,&quot;width&quot;:1456,&quot;resizeWidth&quot;:498,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!aOKM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 424w, https://substackcdn.com/image/fetch/$s_!aOKM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 848w, https://substackcdn.com/image/fetch/$s_!aOKM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!aOKM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff24ddb47-79e7-47a7-ab8e-114a5f7d69c6_2500x2480.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Lunar Libration (<a href="https://www.flickr.com/photos/aucklandmuseum_commons/54960317231/in/album-72177720330692278">source</a>)</figcaption></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.kamilelukosiute.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thank you for reading! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h3>Labs are currently the chokepoint for AI-enabled cyber misuse</h3><p>Currently, the best models with cyber capabilities come from American labs. They are proving themselves useful to attackers, such as to the hacker who <a href="https://www.bloomberg.com/news/articles/2026-02-25/hacker-used-anthropic-s-claude-to-steal-sensitive-mexican-data">used Claude to steal sensitive Mexican government data</a>, and even to Chinese state-sponsored actors who <a href="https://www.anthropic.com/news/disrupting-AI-espionage">use Claude for espionage operations</a>.</p><p>It&#8217;s no longer a question of whether models are useful for attackers; it&#8217;s a question of &#8220;how useful?&#8221; and &#8220;how large are the expected damages?&#8221;</p><p>From a non-proliferation standpoint, the concentration of frontier cyber capabilities in a handful of American labs is structurally useful. When frontier cyber capabilities are held by a handful of companies, the surface area for policy intervention is small: there are only a few entities to regulate, audit, or compel to cooperate with law enforcement. The labs could implement KYC<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> requirements, monitor for misuse, and share threat intelligence &#8211; and <a href="https://openai.com/index/trusted-access-for-cyber/">many do so</a>. Whether or not they are currently exercising this power <em>well</em> is a different question, but the structural chokepoint exists, and it makes intervention possible.</p><p>In practice, though, the frontier lab safety systems for cyber are arguably not very good. The Mexican government case is revealing. The <a href="https://www.bloomberg.com/news/articles/2026-02-25/hacker-used-anthropic-s-claude-to-steal-sensitive-mexican-data">Bloomberg article</a> quotes OpenAI&#8217;s response: &#8220;We have banned the accounts used by this adversary and value the outreach from Gambit Security.&#8221; OpenAI seems to not have caught the misuse themselves through regular monitoring mechanisms, and instead, they were tipped off by an external security firm to investigate.</p><p>Intermediary services complicate monitoring. Even legitimate aggregators like <a href="https://openrouter.ai/">OpenRouter</a> add a layer of indirection &#8211; the model provider sees the aggregator&#8217;s API key, not the end user&#8217;s identity. Illegal reverse proxies, which run on <a href="https://cloud.google.com/blog/topics/threat-intelligence/distillation-experimentation-integration-ai-adversarial-use">stolen API keys and facilitate a black market for unauthorised access</a>, make monitoring essentially impossible. There&#8217;s a case for simply banning<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> intermediary API access that isn&#8217;t directly KYC&#8217;d by the model provider, even if there are legitimate reasons to use aggregators.</p><p>But even if labs were significantly better at detecting and disrupting misuse, the structural concentration that makes it possible is eroding.</p><h3>The open-source capability lag is about 3&#8211;6 months</h3><p>On CyberGym, one of the few public and non-saturated offensive cyber evaluations that I could find, the most capable open-weight model is <a href="https://huggingface.co/zai-org/GLM-5">GLM-5</a>. On this benchmark, the model scores somewhere between Claude Sonnet 4.5 and Claude Opus 4.5. GLM-5 was <a href="https://epoch.ai/benchmarks/eci">released</a> on 11 February 2026; Sonnet 4.5 and Opus 4.5 on 29 September 2025 and 24 November 2025, respectively. The approximate lag in cyber capabilities of GLM-5 to the frontier, closed-weight models is about 3 to 5 months.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pQYu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pQYu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 424w, https://substackcdn.com/image/fetch/$s_!pQYu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 848w, https://substackcdn.com/image/fetch/$s_!pQYu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 1272w, https://substackcdn.com/image/fetch/$s_!pQYu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pQYu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png" width="1456" height="843" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:843,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pQYu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 424w, https://substackcdn.com/image/fetch/$s_!pQYu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 848w, https://substackcdn.com/image/fetch/$s_!pQYu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 1272w, https://substackcdn.com/image/fetch/$s_!pQYu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F354be19a-2542-4e93-8a65-8c214bf3ae11_1600x926.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Data Source: <a href="http://cybergym.io">cybergym.io</a> and <a href="https://epoch.ai/benchmarks/eci">Epoch ECI</a> (for release dates)</em></figcaption></figure></div><p>I hesitate to fully trust this assessment because the CyberGym website collects benchmark scores from the model cards of the original publishers. People have speculated that the Chinese open-source model developers make models less useful than their benchmarks would predict because they optimise entirely for the benchmark scores reported in model cards. Chinese models <a href="https://www.gleech.org/paper">dropped 21%</a> on new AIME 2025 questions versus 10% for Western models, suggesting weaker generalisation to unseen data; and Kimi K2.5&#8217;s reported benchmark scores use <a href="https://huggingface.co/moonshotai/Kimi-K2.5">test-time inflation techniques</a> that don&#8217;t reflect what you&#8217;d get from a single inference pass.</p><p>A better assessment would be done using non-lab, independently run evals. The <a href="https://epoch.ai/benchmarks/eci">Epoch Capabilities Index</a> is a composite evaluation, but it does include in-house run evaluations, including their own <a href="https://epoch.ai/frontiermath">Frontier Math</a> eval. Scores for GLM-5 are not yet available, but the ECI score of Kimi 2.5 (released 27 January 2026) is 148, putting it again between Sonnet 4.5 and Opus 4.5 and indicating a lag of 3&#8211;4.5 months.</p><p>GPT-5 was somewhat of an outlier on the frontier trend, and I&#8217;m not totally sure what to do with it; GLM-5 came out about 6 months after GPT-5. On the whole, the current capabilities lag seems to be about 3&#8211;6 months. The lag matters because there&#8217;s a capability threshold, somewhere around the late-2025 frontier levels, where models go from &#8216;somewhat helpful for cyber&#8217; to &#8216;operationally useful.&#8217; Open-weight models are also approaching this threshold.</p><h3>Once open source is good enough, attackers won&#8217;t need American labs</h3><p>Something happened in December 2025 that even I, someone who primarily writes reports and blog posts for a living, noticed. Andrej Karpathy <a href="https://x.com/karpathy/status/2026731645169185220">explains</a> the step change:</p><blockquote><p>&#8220;There are a number of asterisks but imo coding agents basically didn&#8217;t work before December and basically work since - the models have significantly higher quality, long-term coherence and tenacity and they can power through large and long tasks, well past enough that it is extremely disruptive to the default programming workflow.&#8221;</p></blockquote><p>I don&#8217;t have first-hand evidence, but it seems that a similar step change happened in cyber capabilities as well. In mid-January, Sean Heelan published an excellent <a href="https://sean.heelan.io/2026/01/18/on-the-coming-industrialisation-of-exploit-generation-with-llms/">blog post</a> on using then-frontier models<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> to create exploits. Cybersecurity professionals I&#8217;ve spoken with say the techniques demonstrated in Heelan&#8217;s post would challenge even seasoned practitioners.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> Other cybersecurity professionals tell me that they experienced a shift in how much work agents could do in the past few months as well.</p><p>Once open-weight models cross the same threshold, attackers will switch for two reasons: operational security and operational reliability. Neither of these depend on how well the labs enforce their safety policies.</p><p>A rational attacker has to assume that every API call to a US lab could generate a forensic record on servers that can be subpoenaed by Western law enforcement. Separately, cybercriminal operations are often surprisingly <a href="https://journals.law.harvard.edu/nsj/wp-content/uploads/sites/82/2025/11/Lusthaus_16-Harvard-Natl-Security-J.-2-2025.pdf">industrialised</a>, and building critical infrastructure on top of a service that can revoke your access at any time is an unacceptable operational dependency. Even if open-weight models are behind by a few months, the gains on both fronts will be worth the capability trade-off.</p><p>If the 3&#8211;6 month lag holds and the benchmark scores are real, then open-weight models are right at or just below the &#8220;actually useful for cyber&#8221; threshold now, and will clearly cross it within the next few months. I&#8217;m uncertain as to whether it&#8217;s already happened because Chinese companies are known for eval gaming, and also, I don&#8217;t know anyone actually using these models. Absence of practitioner evidence isn&#8217;t proof it&#8217;s not happening: I just don&#8217;t talk to a lot of people who wouldn&#8217;t just use Codex or Claude.</p><p>There are relatively easy paths for evading Western visibility for attackers by using Chinese APIs. Attackers could steal credit card information to pay for the Moonshot or Z.ai API directly; or just use stolen API keys. Moonshot and Z.ai are required to cooperate with Chinese intelligence, but they have no meaningful incentive or mechanism to report abuse to Western law enforcement. For non-Chinese threat actors (e.g. Russian ransomware groups, Southeast Asian fraud operations), this makes Chinese API access a very low-risk avenue. Chinese state-sponsored actors, meanwhile, presumably don&#8217;t need stolen credit cards at all. Sysdig <a href="https://www.sysdig.com/blog/llmjacking-targets-deepseek">reported that</a> operators who run illegal reverse proxies on stolen API keys added DeepSeek models within days of release, suggesting this shift is already underway.</p><p>For more sophisticated actors, fully private on-prem deployment is also within reach. Deploying Kimi 2.5 on-prem requires significant capex: roughly $200&#8211;500K all in<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>, depending on configuration, for usable agentic coding throughput. But the implicit assumption that this hardware is hard for cybercriminals to get doesn&#8217;t hold up. The <a href="https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0297312">World Cybercrime Index</a> ranks the top six sources of cybercrime as Russia, Ukraine, China, the US, Nigeria, and Romania. Of these, three &#8211; Ukraine, the US, and Romania &#8211; face zero export restrictions on GPU hardware. Russia and China, the two countries most relevant to US national security, are the only top-six sources that face genuine hardware access constraints, and even there, there is a prolific grey market.</p><p>On-prem doesn&#8217;t even have to be the likely path to lose visibility into attacker usage. Between Chinese APIs, non-KYC cloud providers, and compromised infrastructure, there are many ways to run open-weight models outside Western visibility. The appetite for doing so already exists. Mandiant <a href="https://cloud.google.com/blog/topics/threat-intelligence/distillation-experimentation-integration-ai-adversarial-use">reports</a> that &#8220;there is an enduring market for AI services specifically designed to support malicious activity&#8221; and I see no reason why this market will go away. One such service, Xanthorox, advertises itself as running &#8220;security on our dedicated servers&#8221; and &#8220;not sharing data with third-party corporations,&#8221; showing that this is clearly a desirable quality for attackers<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>.</p><h3>Distillation might be accelerating the proliferation of offensive cyber capabilities</h3><p>Anthropic <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">recently accused</a> Moonshot, the makers of Kimi, as well as DeepSeek and Minimax, of illicitly extracting Claude&#8217;s capabilities to train their own models. It was a large operation: the Chinese labs &#8220;generated over 16 million exchanges with Claude through approximately 24,000 fraudulent accounts.&#8221; Google Threat Intelligence also <a href="https://cloud.google.com/blog/topics/threat-intelligence/distillation-experimentation-integration-ai-adversarial-use">wrote in the past few weeks</a> of a distillation attack campaign against Gemini, identifying &#8220;over 100,000 prompts.&#8221;</p><p>Moonshot is the same company that makes Kimi 2.5, the frontier-on-benchmarks-model, whose API I previously described above as a low-risk access point for attackers wanting frontier capabilities without Western law enforcement oversight. Moonshot distills from Claude and makes these capabilities freely available to adversaries and attackers. The case for preventing distillation is larger than just protecting corporate IP. If distillation is what&#8217;s keeping the open-weight lag to just 3&#8211;6 months, then anti-distillation efforts are also a cybersecurity misuse intervention. Mitigating distillation directly protects the window in which labs can still see and respond to misuse.</p><p>To what extent are Kimi and GLM&#8217;s cyber capabilities due to distillation? It&#8217;s hard to say, but if these models are only 3&#8211;6 months behind because they&#8217;re collecting data from Anthropic, OpenAI, and Google DeepMind and training on it, then making distillation harder is one of the highest-leverage interventions available right now.</p><h3>Plan for resilience, not for control</h3><p>Even if we can fully protect against distillation attacks, though, it won&#8217;t fully solve the problem. The Chinese state-sponsored espionage campaign that Anthropic disrupted in September 2025 was conducted using Claude Code, which at the time ran on Opus 4.1. Current open-weight models already score above Opus 4.1 on capability benchmarks. The capabilities that enabled the first documented autonomous cyber espionage operation are, in all likelihood, already available in the near-frontier open-weight models. There&#8217;s also probably an elicitation overhang: not every attacker has yet discovered the best prompting strategies and tool-use patterns for offensive cyber, but they will.</p><p>We have to reckon with the fact that we now live in a world where attackers have access to advanced cyber-attack capabilities. There is no putting the cat back in the bag. Even if we expect no further progress from open or closed models, we should expect future harm as attackers learn how to best apply what&#8217;s already available.</p><p>The lab-based safety paradigm buys time, but not much, and anti-distillation efforts can extend it but can&#8217;t preserve it indefinitely. The priority now should be investing in defences that work regardless of where the model lives &#8211; hardening the institutions most likely to be targeted, building AI-powered defensive tooling <a href="https://kamilelukosiute.com/posts/design-for-the-defenders-you-care-about-or-risk-being-useless.html">designed for real operational constraints</a>, and developing detection and attribution capabilities that don&#8217;t depend on visibility into model providers.</p><p>In my next blog post, I&#8217;ll describe what a serious investment agenda along these lines would look like.</p><p></p><p><em>Thank you to Adam Swanda, Jake Steckler, Amelia Michael, Aidan Homewood, Matthew van der Merwe, and Alan Chan for feedback on this piece.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Know Your Customer &#8211; requiring users to identify themselves, e.g. by creating an account, providing a credit card, or verifying identity. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I don&#8217;t love that my best policy analogy here is an authoritarian government, but: China <a href="https://www.weforum.org/stories/2022/01/what-s-behind-china-s-cryptocurrency-ban/">banned Bitcoin in 2021</a>, on the basis of the same reasoning: legitimate use cases exist, sure, but the crime-enabling downside isn&#8217;t worth it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Opus 4.5 and GPT-5.2</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Shout out to Heron Security for <a href="https://www.linkedin.com/posts/eli-parkes_moneys-not-the-bottleneck-talent-is-if-activity-7430522486296965120-mMFu?utm_source=share&amp;utm_medium=member_desktop&amp;rcm=ACoAAB5GEqIB39qcLLKPYBip3vqzvEKG6Wv4zDU">hosting a delegation</a> of 14 cybersecurity professionals who participated in my workshop on AI cyber capabilities.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Running Kimi requires 4&#8211;8 high-end NVIDIA GPUs (e.g. H100, B200), with total hardware costs ranging from roughly $120,000 to $300,000 depending on configuration. Individual B200 GPUs sell for $30,000&#8211;$40,000 each (Epoch AI, <a href="https://epoch.ai/data-insights/b200-cost-breakdown">Dec 2025</a>; <a href="https://epoch.ai/data/ai-chip-sales-documentation">Jan 2026</a>); an 8-GPU H100 server runs ~$190,000 at hyperscaler pricing or ~$280,000 at retail (Epoch AI, <a href="https://epoch.ai/data/ai-chip-sales-documentation">Jan 2026</a>). Assuming GPU hardware is roughly 60% of total datacenter operating costs, total deployment costs are approximately $200,000&#8211;$500,000. This hardware setup would support roughly 10&#8211;20 concurrent agentic sessions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>This Xanthorox is a bit of a meme because Mandiant reports that Xanthorox is not a custom AI at all but rather using stolen API keys for frontier models and routing traffic to them; but I expect more advanced attackers to do better and actually maintain proper opsec.</p></div></div>]]></content:encoded></item><item><title><![CDATA[American Character]]></title><description><![CDATA[Originally published September 1, 2023; cross-posted to Substack on February 7, 2026]]></description><link>https://blog.kamilelukosiute.com/p/american-character</link><guid isPermaLink="false">https://blog.kamilelukosiute.com/p/american-character</guid><dc:creator><![CDATA[Kamile Lukosiute]]></dc:creator><pubDate>Sat, 07 Feb 2026 20:12:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2ei3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Originally published September 1, 2023; cross-posted to Substack on February 7, 2026</em></p><p>A few weeks ago, I had some spare time, so I went for a long walk. In the end, I walked for twelve days and 200 miles of the Pacific Crest Trail. Rookie numbers compared to those who walk the full trail (which is 2600 miles in length total), but given that this was my first long distance hike, I think I did well.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2ei3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2ei3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 424w, https://substackcdn.com/image/fetch/$s_!2ei3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 848w, https://substackcdn.com/image/fetch/$s_!2ei3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 1272w, https://substackcdn.com/image/fetch/$s_!2ei3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2ei3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png" width="1000" height="667" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa9f95c3-4bad-417b-a429-107200772758_1000x667.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:667,&quot;width&quot;:1000,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1074071,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://kamilelukosiute.substack.com/i/187229932?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2ei3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 424w, https://substackcdn.com/image/fetch/$s_!2ei3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 848w, https://substackcdn.com/image/fetch/$s_!2ei3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 1272w, https://substackcdn.com/image/fetch/$s_!2ei3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa9f95c3-4bad-417b-a429-107200772758_1000x667.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In those twelve days, I pushed myself out of my comfort zone constantly - both physically and mentally. I started by hiking 10 or so miles a day until I acclimated to the altitude and the strenuous activity, and by the end of the trip, I was walking 21-23 miles a day. The physical challenge was not the biggest, however. No, the biggest challenge was the mental one of being constantly alone.</p><p>Most people who hike the PCT start at the Mexican border or Canada and end up members of a &#8220;Bubble&#8221; - a dense cluster of hikers, separated only by a few miles between them. At night, they tend to cluster at water sources or known campsites. When you hike the PCT in full, you&#8217;re never really alone for all that long. If you pick an arbitrary section of the PCT to hike at an arbitrary time (like I did), well, there&#8217;s no guarantee there will be anyone else on trail.</p><p>For most of the eleven nights I was hiking, I camped completely alone. Sometimes twenty four hours would go by until I saw another human being. This might lovely to someone sitting at home, but it feels terrifying in the wilderness. You may be 30 miles from the nearest semblance of civilization, with only your water filter and homemade first aid kit to save you in case of emergency. The first few hours are relaxing, peaceful, but after a while, all the likely and unlikely scenarios start running through your head. You realize how little knowledge you truly have about surviving in the wilderness. How many venemous snakes are there in the Sierras? How much tree cover is sufficient to protect you in a lightning storm on top of a mountain ridge? What if there&#8217;s a bear with rabies? Can bears get rabies? What if grizzly bears weren&#8217;t completely hunted into extinction in the 1800s and I&#8217;m going to be attacked by the lone survivor?</p><p>Rational thinking turns irrational when there is not a single soul around to keep you grounded. Your own thoughts keep you awake, afraid, tired, sleepless. You long for a person to talk to, any person.</p><p>It was in this mental state, after hiking twelve days straight, after hiking about 23 miles on my last day, that I arrived at the intersection of the PCT and Big Creek Road in Bucks Lake, California. My map told me that this was a spot that one could hitchhike from into the town of Quincy. I had been dreaming about the prospect of a shower and a clean bed in a motel from the moment I started hiking that morning.</p><p>I arrived at aforementioned road at 6:45pm and quickly realized that it was a rarely traversed country road. I stood there for 30 or 40 minutes, but <em>not a single car drove by.</em> The sun was setting, and I was worried I would have to for one last time camp completely alone - along the side of this road. I was getting somewhat desperate, when I saw the first car: a red Jeep, older model, the kind meant for off-roading. I stick out my thumb, they stop. Two men, late 20s, wearing camo, carrying hunting bows. I ask if they would take me to town, they say sure, but we&#8217;re gonna be hunting until night falls in a few hours, then we&#8217;ll pick you up here... unless you wanna go deer hunting with us.</p><p>Perhaps I shouldn&#8217;t have been so trusting - perhaps I should have been afraid of being kidnapped, but I was not. In that moment, I was so excited to see real people, I was so traumatized by being alone, I was so thankful that they had stopped to pick me up. I had never been deer hunting before, but I said, you you know what, I&#8217;d love to join you for an hour or two of deer hunting. My other option was to stand there, wait for the next car that may never come as it got darker and darker.</p><p>So I hopped into the back of the Jeep, we drove up the mountain (that I had just climbed down), braking only for deer we spotted. They told me about their hometown, their friends, their jobs as power line repairmen (working predominantly in winter when the storms break the lines), the injuries they sustain on the job. They told me about how they hunt for bears and deer and waterfowl, how their freezers full of game meat make them feel prepared for emergency and self-sufficient. As we rode in the Jeep, they offered me a beer, which I gladly accepted (Coors Light may as well be the nectar of the gods after 13 hours of hiking).</p><p>In my daily life, I don&#8217;t encounter people who regularly hunt and fish, who drive pickup trucks and Harley&#8217;s, who own guns, who didn&#8217;t go to college. I live in my own, self-selected little bubble of computer scientists, physicists, the occasional biologist. So this chance encounter was a fascinating glimpse into another world, their world, one they were very open to introduce me to. We talked a lot while trying to spot deer. They were, to use their own word - hillbillies - but they were also intelligent, self-reflective, kind, and thoughtful. When they asked me what I was doing out in the backcountry, alone, hiking, I told them the truth - I lost my dream job a few months ago, and it still really hurt me, and I still hated myself for messing it all up, so I wanted to think about it. One of the guys turned around and said, &#8220;Was it your dream job or a job at your dream company?&#8221; He fully understood what I was going through, he fully understood what had happened to me, he just got it and I didn&#8217;t need to explain.</p><p>After 2 hours, we had spotted a few does, but alas California only gives out tags for bucks at this time of year, so we ended the hunt empty handed. They drove me into town, we all went to Safeway for snacks, and then they drove me to a motel. I went to ask if there were rooms available, while the two dudes waited for me in the car. No rooms. They drove me to another hotel; again, no rooms. Finally, it was decided that it was best for me to just go to a campground where there would certainly be space, so they brought me there. These concealed-carrying, deer-hunting, self-proclaimed hillbillies showed me nothing but respect, kindness, care, patience, generosity, curiosity and expected nothing in return.</p><p>I used to think that you needed to have a certain type of life to develop the character traits that I now value the most - that you needed to be well-read and educated to become curious, generous, tenacious, dedicated, kind. I rarely saw these traits that I now value so much growing up, and I always assumed it was because poor, uneducated, immigrants (like myself) don&#8217;t develop character traits like this without extraordinary amounts of effort (and favorable circumstance). But this was not true; these character traits <em>can</em> be found everywhere, and perhaps they are as likely to be found among the well-read, educated tech folk that I associate with now as they are in my immigrant community. But these traits are <em>rare</em>, incredibly rare. These character traits are the best America has to offer - and the best is rare. If you&#8217;re lucky, you&#8217;ll stick your thumb out hitchhiking and run into the best on a quiet, rarely traversed country road.</p>]]></content:encoded></item></channel></rss>