<p>started listening <a href="https://neodb.social/search?r=1&q=https://neodb.social/podcast/episode/0cum85D7IcleHZ36OQT4q9" rel="nofollow">The Tucker Carlson Show - Sam Altman on God, Elon Musk and the Mysterious Death of His Former Employee</a><br>absolutely epic episode<br><a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/ai/" rel="tag">#AI</a></p><p><a href="/tags/neodb/" rel="tag">#neodb</a></p>
llm
<p>iPadを何とかできるのは、うちの艦長くらいですよ</p><p>Apple's Cheap MacBook: What to Expect in 2026 <a href="https://www.macrumors.com/2025/11/07/low-cost-macbook-rumors-2026/" rel="nofollow" class="ellipsis" title="www.macrumors.com/2025/11/07/low-cost-macbook-rumors-2026/"><span class="invisible">https://</span><span class="ellipsis">www.macrumors.com/2025/11/07/l</span><span class="invisible">ow-cost-macbook-rumors-2026/</span></a></p><p><a href="/tags/apple/" rel="tag">#Apple</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/news/" rel="tag">#news</a> <a href="/tags/bot/" rel="tag">#bot</a></p>
Since I'm job and work hunting I tend to see the absurd new job titles that are bouncing around in the tech sector. The latest, which I've seen twice today, is "artificial general intelligence engineer" or some permutation thereof. I do my best to spend the minimum possible time on these and have no guess about whether they're legitimate.<br><br><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/agi/" rel="tag">#AGI</a><br>
I put the text below on LinkedIn in response to a post there and figured I'd share it here too because it's a bit of a step from what I've been posting previously on this topic and might be of some use to someone.<br><br>In retrospect I might have written non-sense in place of nonsense.<br><br>If you're in tech the Han reference might be a bit out of your comfort zone, but Andrews is accessible and measured.<br><br><br><p>It's nonsense to say that coding will be replaced with "good judgment". There's a presupposition behind that, a worldview, that can't possibly fly. It's sometimes called the theory-free ideal: given enough data, we don't need theory to understand the world. It surfaces in AI/LLM/programming rhetoric in the form that we don't need to code anymore because LLM's can do most of it. Programming is a form of theory-building (and understanding), while LLMs are vast fuzzy data store and retrieval systems, so the theory-free ideal dictates the latter can/should replace the former. But it only takes a moment's reflection to see that nothing, let alone programming, can be theory-free; it's a kind of "view from nowhere" way of thinking, an attempt to resurrect Laplace's demon that ignores everything we've learned in the >200 years since Laplace forwarded that idea. In that respect it's a (neo)reactionary viewpoint, and it's maybe not a coincidence that people with neoreactionary politics tend to hold it. Anyone who needs a more formal argument can read Mel Andrews's The Immortal Science of ML: Machine Learning & the Theory-Free Ideal, or Byung-Chul Han's Psychopolitics (which argues, among other things, that this is a nihilistic).<br></p><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/coding/" rel="tag">#coding</a> <a href="/tags/dev/" rel="tag">#dev</a> <a href="/tags/tech/" rel="tag">#tech</a> <a href="/tags/softwaredevelopment/" rel="tag">#SoftwareDevelopment</a> <a href="/tags/programming/" rel="tag">#programming</a> <a href="/tags/nihilism/" rel="tag">#nihilism</a> <a href="/tags/linkedin/" rel="tag">#LinkedIn</a><br>
<p>We're going to put together a resource for musicians on how they can have album art without using large language models. Do you have any resources you'd like to see in that? Free image libraries? Friends who can help do design? Basic how-tos?</p><p><a href="/tags/mutualaid/" rel="tag">#mutualaid</a> <a href="/tags/solidarity/" rel="tag">#solidarity</a> <a href="/tags/helpingeachother/" rel="tag">#helpingeachother</a> <a href="/tags/mirlo/" rel="tag">#mirlo</a> <a href="/tags/resources/" rel="tag">#resources</a> <a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/llm/" rel="tag">#llm</a></p>
<p><span class="h-card"><a href="https://ovo.st/club/poll" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>poll</span></a></span> <span class="h-card"><a href="https://ovo.st/club/board" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>board</span></a></span> <br>你觉得哪家公司LLM的模型最好?</p><p><a href="/tags/无聊投票/" rel="tag">#无聊投票</a> <a href="/tags/llm/" rel="tag">#LLM</a></p>
<div class="poll">
<h3 style="display: none;">Options: <small>(choose one)</small></h3>
<ul>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="11 votes">14%</span>
<span class="poll-option-text">Anthropic</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="28 votes">36%</span>
<span class="poll-option-text">Google</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="8 votes">10%</span>
<span class="poll-option-text">OpenAI</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="1 votes">1%</span>
<span class="poll-option-text">xAI</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="0 votes">0%</span>
<span class="poll-option-text">深度求索(DeepSeek)</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="0 votes">0%</span>
<span class="poll-option-text">月之暗面(Moonshot)</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="1 votes">1%</span>
<span class="poll-option-text">阿里巴巴(Alibaba)</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="0 votes">0%</span>
<span class="poll-option-text">百度(Baidu)</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="0 votes">0%</span>
<span class="poll-option-text">其它(请留言说明)</span>
</label>
</li>
<li>
<label class="poll-option">
<input style="display:none" name="vote-options" type="radio" value="0">
<span class="poll-number" title="28 votes">36%</span>
<span class="poll-option-text">围观通道</span>
</label>
</li>
</ul>
<div class="poll-footer">
<span class="vote-total">77 votes</span>
—
<span class="vote-end">Ended 55d ago</span>
<span class="todo">Polls are currently display only</span>
</div>
</div>
<p><a href="/tags/lispygopherclimate/" rel="tag">#lispyGopherClimate</a> <a href="/tags/lisp/" rel="tag">#lisp</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/podcast/" rel="tag">#podcast</a> <a href="/tags/archive/" rel="tag">#archive</a>, <a href="/tags/climate/" rel="tag">#climate</a> <a href="/tags/haiku/" rel="tag">#haiku</a> by <span class="h-card"><a href="https://climatejustice.social/@kentpitman" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>kentpitman</span></a></span> <br><a href="https://communitymedia.video/w/c3GdAXe7BQTbK3VrcXCm7E" rel="nofollow" class="ellipsis" title="communitymedia.video/w/c3GdAXe7BQTbK3VrcXCm7E"><span class="invisible">https://</span><span class="ellipsis">communitymedia.video/w/c3GdAXe</span><span class="invisible">7BQTbK3VrcXCm7E</span></a><br>& <span class="h-card"><a href="https://fe.disroot.org/users/ramin_hal9001" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>ramin_hal9001</span></a></span> <br>On the <a href="/tags/climate/" rel="tag">#climate</a> I would like to talk about the company that found <a href="/tags/curl/" rel="tag">#curl</a> and <a href="/tags/openssl/" rel="tag">#openssl</a>'s <a href="/tags/deeplearning/" rel="tag">#deeplearning</a> many (10ish) 0-day vulns "using <a href="/tags/ai/" rel="tag">#ai</a> ". (<a href="/tags/llm/" rel="tag">#llm</a> s were involved).</p><p>This obviously relates to my <a href="/tags/lisp/" rel="tag">#lisp</a> <a href="/tags/symbolic/" rel="tag">#symbolic</a> <a href="/tags/dl/" rel="tag">#DL</a> <a href="https://screwlisp.small-web.org/conditions/symbolic-d-l/" rel="nofollow" class="ellipsis" title="screwlisp.small-web.org/conditions/symbolic-d-l/"><span class="invisible">https://</span><span class="ellipsis">screwlisp.small-web.org/condit</span><span class="invisible">ions/symbolic-d-l/</span></a> (ffnn equiv). Thanks to everyone involved with that so far.</p><p>I implemented that using <a href="/tags/commonlisp/" rel="tag">#commonLisp</a> <a href="/tags/condition/" rel="tag">#condition</a> handling viz KMP.</p>
Edited 61d ago
Compare and contrast<br><br>This:<br><p>In the year of the city 2274, the remnants of human civilization live in a sealed city beneath a cluster of geodesic domes, a utopia run by computer. The citizens live a hedonistic lifestyle, but when they turn 30 must enter the "Carrousel", a public ritual that destroys their bodies, under the pretense they would be "Renewed" or reborn.<br></p>(<a href="https://en.wikipedia.org/wiki/Logan's_Run_(film)" rel="nofollow">Logans Run</a>)<br><br>and this:<br><p>In the year of the city 2274, the colony of human beings on Mars live in a sealed city beneath a cluster of geodesic domes, a utopia run by generative AI. The citizens live a hedonistic lifestyle, but when they turn 30 must enter the "Cloud", a public ritual that destroys their bodies, under the pretense their consciousness would be uploaded to a computer and live forever.<br></p><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/mars/" rel="tag">#Mars</a> <a href="/tags/eugenics/" rel="tag">#eugenics</a> <a href="/tags/logansrun/" rel="tag">#LogansRun</a> <a href="/tags/sciencefiction/" rel="tag">#ScienceFiction</a> <a href="/tags/dystopia/" rel="tag">#dystopia</a><br>
<p><a href="/tags/lispygopherclimate/" rel="tag">#lispyGopherClimate</a> <a href="/tags/live/" rel="tag">#live</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/podcast/" rel="tag">#podcast</a> hosted by <span class="h-card"><a href="https://fe.disroot.org/users/ramin_hal9001" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>ramin_hal9001</span></a></span> <a href="https://tilde.town/~ramin_hal9001/articles/ai-cannot-live-up-to-the-hype.html" rel="nofollow" class="ellipsis" title="tilde.town/~ramin_hal9001/articles/ai-cannot-live-up-to-the-hype.html"><span class="invisible">https://</span><span class="ellipsis">tilde.town/~ramin_hal9001/arti</span><span class="invisible">cles/ai-cannot-live-up-to-the-hype.html</span></a> this week!</p><p><a href="https://communitymedia.video/w/sBNPeWFJ7NuDVCjkAv7fkX" rel="nofollow" class="ellipsis" title="communitymedia.video/w/sBNPeWFJ7NuDVCjkAv7fkX"><span class="invisible">https://</span><span class="ellipsis">communitymedia.video/w/sBNPeWF</span><span class="invisible">J7NuDVCjkAv7fkX</span></a></p><p>My guess as to the topic from the main show toot which is </p><p><a href="https://fe.disroot.org/objects/88f34eba-ccaf-407e-8505-8a2a3bd1613b" rel="nofollow" class="ellipsis" title="fe.disroot.org/objects/88f34eba-ccaf-407e-8505-8a2a3bd1613b"><span class="invisible">https://</span><span class="ellipsis">fe.disroot.org/objects/88f34eb</span><span class="invisible">a-ccaf-407e-8505-8a2a3bd1613b</span></a> <- visit</p><p><span class="h-card"><a href="https://climatejustice.social/@kentpitman" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>kentpitman</span></a></span> 's poem</p><p>Repudiating the <a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/llm/" rel="tag">#llm</a> hype!</p><p>And <a href="/tags/environmental/" rel="tag">#environmental</a> damage <a href="/tags/climate/" rel="tag">#climate</a> </p><p><a href="/tags/commonlisp/" rel="tag">#commonLisp</a> <a href="/tags/lisp/" rel="tag">#lisp</a> <a href="/tags/typetheory/" rel="tag">#typeTheory</a> <a href="/tags/coalton/" rel="tag">#coalton</a> common lisp static typing DSL <a href="https://coalton-lang.github.io/" rel="nofollow"><span class="invisible">https://</span>coalton-lang.github.io/</a> ,<br><a href="/tags/els/" rel="tag">#ELS</a> 2025 talk by Robert Smith. Robert emailed a note.</p><p><a href="/tags/programming/" rel="tag">#programming</a></p>
Edited 54d ago
<p>Good AI research should tell us something about life, or it should help people. I hate seeing research about automating what people do. It's not a good goal for science or society! I was recently reminded of this by a paper applying LLMs to math.</p><p>This domain has many good questions: what do we mean when we say a person "solves math problems"? What are they actually doing? How is this like or not like what an LLM does? How might mathematicians benefit from this?</p><p>Instead, we get papers that pit an LLM against a human on a math problems dataset. This is great for claiming "AI has superhuman math abilities now!", but it's debatable whether good answers in a test-taking environment have anything to do with logic, reasoning, or creative problem solving. Instead of exploring to what extent LLMs are "really intelligent" vs. "stochastic parrots" (and perhaps the same question for humans), it reduces everything down to a number, one that hides the deeper problem and seems far more definitive than it is.<br><a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/llm/" rel="tag">#llm</a></p>
Edited 50d ago
<p>I hate OpenAI but I had to use Whisper to help someone make accessible content. I hate that I had to use Whisper to do it because it comes from OpenAI.</p><p>But I don't know of any other way to get a text transcription from a media file that is free/open. (Besides doing it manually.)</p><p>I tell myself because it's for education and accessibility it's okay, but I still don't like it.</p><p><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/openai/" rel="tag">#OpenAI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/whisper/" rel="tag">#whisper</a></p>
<p><a href="/tags/til/" rel="tag">#TIL</a> that HDD prices also increased (because of "AI") like the SSD/memory prices before. I was slowly building my home NAS, thinking that if I use HDD, not SSD, then I should race only with the possibility of Internet shutdown in my country and hope that I'll build the machine before the Internet will become unusable or completely turned off, and I will be able to preserve at least something from human knowledge and creativity. But … looks like for now I should race not only with censorship, but also with fucking "AI" corporations <img src="https://neodb.social/media/emoji/bsd.cafe/drgn_knife_angry.png" class="emoji" alt=":drgn_knife_angry:" title=":drgn_knife_angry:"></p><p>The worst timeline ever, I never saw how the price of something was decreased in my life. When I moved to my city at 2008, the bus fare was near 18 roubles. For now it is 88 roubles — the 389% rise. Fuck this shit.</p><p>I think, one time the novel "Walkaway" from Cory Doctorow will become not a novel but a real-life. Because there are no mass-adopted solution for now — you could choose only in between "you own nothing and be happy" and "you own nothing and welcome to GULAG" <img src="https://neodb.social/media/emoji/bsd.cafe/drgn_roar_angry.png" class="emoji" alt=":drgn_roar_angry:" title=":drgn_roar_angry:">. So the good enough solution — not to participate in that circus at all and walk away.</p><p><a href="/tags/rant/" rel="tag">#rant</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/hdd/" rel="tag">#HDD</a></p>
OpenClaw founder Steinberger joins OpenAI, open-source bot becomes foundation<br><br>From <a href="https://www.reuters.com/business/openclaw-founder-steinberger-joins-openai-open-source-bot-becomes-foundation-2026-02-15/" rel="nofollow" class="ellipsis" title="www.reuters.com/business/openclaw-founder-steinberger-joins-openai-open-source-bot-becomes-foundation-2026-02-15/"><span class="invisible">https://</span><span class="ellipsis">www.reuters.com/business/openc</span><span class="invisible">law-founder-steinberger-joins-openai-open-source-bot-becomes-foundation-2026-02-15/</span></a><br><br>Everything I've read about OpenClaw suggests it's the NFT of AI. These folks need the fiction that AI is approaching "consciousness", or at least "agency", to continue.<br><br><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/agenticai/" rel="tag">#AgenticAI</a> <a href="/tags/vibecoding/" rel="tag">#VibeCoding</a> <a href="/tags/openai/" rel="tag">#OpenAI</a> <a href="/tags/openclaw/" rel="tag">#OpenClaw</a><br>
<p>看到这句😂</p><p>我用象棋软件或者围棋软件可以下赢任何一个人类的世界冠军,但我不会满大街的说我是下棋高手。<a href="/tags/llm/" rel="tag">#LLM</a><br><a href="https://t.me/hyi0618/10750" rel="nofollow"><span class="invisible">https://</span>t.me/hyi0618/10750</a></p>
<p>I'm so confused. I just found a small body of literature applying LLMs to reinforcement learning type tasks, exploring the use of LLMs for "autonomous decision making."</p><p>I guess people are building more LLM agent systems, and we ought to understand them and what makes them better / worse at what they do.</p><p>But I still feel like LLMs are fundamentally not suited to decision making tasks. They don't weigh options and decide. At best, you could say they interpolate what a reasonable choice might look like based on the examples of people making choices in their training data.</p><p>That's... really not the same thing! Like, not at all. It's impressive that this sometimes works, but this seems very silly to me when we could be using actual RL systems that really are making informed decisions from experience, with mathematical rigor to estimate the quality of those choices.</p><p><a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/rl/" rel="tag">#rl</a></p>
<p>〈Over 60% of Asia-Pacific enterprises plan to increase sovereign AI spending〉 </p><p><a href="https://www.bworldonline.com/technology/2026/03/05/734185/over-60-of-asia-pacific-enterprises-plan-to-increase-sovereign-ai-spending/" rel="nofollow" class="ellipsis" title="www.bworldonline.com/technology/2026/03/05/734185/over-60-of-asia-pacific-enterprises-plan-to-increase-sovereign-ai-spending/"><span class="invisible">https://</span><span class="ellipsis">www.bworldonline.com/technolog</span><span class="invisible">y/2026/03/05/734185/over-60-of-asia-pacific-enterprises-plan-to-increase-sovereign-ai-spending/</span></a></p><p><a href="/tags/sovereignai/" rel="tag">#SovereignAI</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/asiapacific/" rel="tag">#AsiaPacific</a></p>
Something I've learned from Ruth Ben-Ghiat: aspiring authoritarians purposely engineer situations in which people are invited to give up their values and morals and make decisions that compromise their sense of right and wrong. Moral decay, moral injury, and subsequently moral collapse become so intolerable that afflicted people will blame anything else but their own choices or the leader they threw in with, which of course are the only proximate causes it would be helpful to implicate. The more compromising decisions they make, the more they are drawn into the authoritarian's orbit.<br><br>There is no question that it is indefensible to use generative AI systems as they are currently constituted, especially the commercial ones, once one becomes aware of how they are made and operated and the destructive consequences they have already had and will surely continue to have. Among the many reasons using these tools is indefensible is that they represent an authoritarian invitation. You're invited to trade your morals and ethics for a bit of convenience, a reduction in friction, a learning experience, a rhetorical flourish, or maybe (a kind of) status. You thereby align yourself more and more with people who say things like "water is fake" or "fuck earth" as they make the computer systems enabling the horrors we're watching unfold on social media. You start to tell yourself stories, complexifying stories that explain why it's OK you did this thing that you know is not OK. You move in the direction of people who are already telling themselves stories like this. Maybe their stories have superior analgesic qualities to yours.<br><br>Nobody needs to go down this path.<br><br><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/ethics/" rel="tag">#ethics</a> <a href="/tags/morality/" rel="tag">#morality</a> <a href="/tags/authoritarianism/" rel="tag">#authoritarianism</a><br>
<p>A lab mate shared this write up of Don Knuth using LLMs to solve a math problem: <a href="https://www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf" rel="nofollow" class="ellipsis" title="www-cs-faculty.stanford.edu/~knuth/papers/claude-cycles.pdf"><span class="invisible">https://</span><span class="ellipsis">www-cs-faculty.stanford.edu/~k</span><span class="invisible">nuth/papers/claude-cycles.pdf</span></a></p><p>It's clear that using Claude did help them arrive at some new understanding here, which is wonderful. I'm happy for them.</p><p>However, I'm upset by how much they personify Claude and attribute the solution to "him."</p><p>From this narrative, it's clear that the humans were very actively involved from beginning to end. Claude was a helpful tool, but it did not solve this problem on its own. What role did it actually play? How was it like or unlike a human collaborator on this problem?</p><p>It did generate a crucial insight, but where did that come from? Was it plagiarized from some unknown source? Did it "just emerge" from text completion and interpolation in latent space? Do we need some other explanation for Claude's apparent creativity?</p><p>These folks don't care. They just wanted a solution, which they attribute to Claude, and leave it at that. I think that's a serious problem.</p><p><a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/science/" rel="tag">#science</a> <a href="/tags/math/" rel="tag">#math</a></p>
<p>An open letter to <a href="/tags/grammarly/" rel="tag">#Grammarly</a> and other plagiarists, thieves and slop merchants<br>by Maureen Ryan</p><p>> You do not get to say that you don’t understand these very basic issues of autonomy, respect and self-control. You. Do. Not.</p><p><a href="/tags/maureenryan/" rel="tag">#MaureenRyan</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/slop/" rel="tag">#Slop</a> <a href="/tags/llm/" rel="tag">#LLM</a> <br><a href="/tags/identitytheft/" rel="tag">#IdentityTheft</a> <br><a href="https://www.moryan.com/an-open-letter-to-grammarly-and-other-plagiarists-thieves-and-slop-merchants/" rel="nofollow" class="ellipsis" title="www.moryan.com/an-open-letter-to-grammarly-and-other-plagiarists-thieves-and-slop-merchants/"><span class="invisible">https://</span><span class="ellipsis">www.moryan.com/an-open-letter-</span><span class="invisible">to-grammarly-and-other-plagiarists-thieves-and-slop-merchants/</span></a></p>
<p>Your <a href="/tags/llm/" rel="tag">#LLM</a> Doesn't Write Correct Code. It Writes Plausible Code.</p><p>Deployment is an act of faith, not engineering.<br> <br>[…] THIS is the failure mode. Not broken syntax or missing semicolons. The code is syntactically and semantically correct. It does what was asked for. It just does not do what the situation requires. In the SQLite case, the intent was “implement a query planner” and the result is a query planner that plans every query as a full table scan. In the disk daemon case, the intent was “manage disk space intelligently” and the result is 82,000 lines of intelligence applied to a problem that needs none. Both projects fulfill the prompt. Neither solves the problem.<br><a href="https://blog.katanaquant.com/p/your-llm-doesnt-write-correct-code" rel="nofollow" class="ellipsis" title="blog.katanaquant.com/p/your-llm-doesnt-write-correct-code"><span class="invisible">https://</span><span class="ellipsis">blog.katanaquant.com/p/your-ll</span><span class="invisible">m-doesnt-write-correct-code</span></a></p><p><a href="/tags/aihype/" rel="tag">#aihype</a> <a href="/tags/theaicon/" rel="tag">#theaicon</a></p>
<p>Boost plz!</p><p>Looking for critical scholarship on the use of "AI" by library/archive workers. University libraries in particular, but adjacent and tangentially-relevant-at-best stuff is welcome too. Any format is fine: books, papers, blogposts, whatever. If it's good, gimme all you've got!</p><p>Looks like we're gonna have a department-wide conversation about people using LLMs, and it's being framed as "we're all using it, but we're not talking about it, so let's make sure we're all on the same page about using it responsibly" ... I'll of course be pushing the "there's basically no way to use it responsibly" position, and I'd like to arm myself and others with some critical analyses of issues related to its use in library/archive spaces.</p><p><a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/libraries/" rel="tag">#libraries</a> <a href="/tags/archives/" rel="tag">#archives</a></p>
Edited 12d ago
<p>The whole LLM as a service business model has a fundamental flaw to it. The cost of operating the data centres is an order of magnitude higher than the profit.</p><p>But if models get efficient enough to bring the costs down, then they become efficient enough to run locally. So, either it’s too expensive to operate, or nobody will want to use it as a service because running your own gives you privacy and flexibility.</p><p>The fact that investors don't get this is frankly incredible.</p><p><a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/economy/" rel="tag">#economy</a></p>
<p>I had to get this idea out of my head. <a href="/tags/theylive/" rel="tag">#TheyLive</a> <a href="/tags/llm/" rel="tag">#LLM</a></p>
<p>一些 <a href="/tags/阴谋论/" rel="tag">#阴谋论</a> 在中文互联网上流传,认为一些失踪人口的意识被上传,以用于开发豆包等大语言模型。</p><p><a href="https://www.zhihu.com/question/2022623696783774161" rel="nofollow" class="ellipsis" title="www.zhihu.com/question/2022623696783774161"><span class="invisible">https://</span><span class="ellipsis">www.zhihu.com/question/2022623</span><span class="invisible">696783774161</span></a><br><a href="https://xcancel.com/YesterdayOrange/status/2038891964415037591" rel="nofollow" class="ellipsis" title="xcancel.com/YesterdayOrange/status/2038891964415037591"><span class="invisible">https://</span><span class="ellipsis">xcancel.com/YesterdayOrange/st</span><span class="invisible">atus/2038891964415037591</span></a></p><p><a href="/tags/conspiracy/" rel="tag">#conspiracy</a> <a href="/tags/internetmysteries/" rel="tag">#internetmysteries</a> <a href="/tags/missingperson/" rel="tag">#missingperson</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/互联网谜团/" rel="tag">#互联网谜团</a> <a href="/tags/失踪人口/" rel="tag">#失踪人口</a> <a href="/tags/人工智能/" rel="tag">#人工智能</a> <a href="/tags/大语言模型/" rel="tag">#大语言模型</a> <a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/llms/" rel="tag">#llms</a></p>
Edited 5d ago
<p>DAIR is a research institute that is highly sceptical about AI hype and the big tech companies behind it. You can follow their excellent video account at:</p><p>➡️ <span class="h-card"><a href="https://peertube.dair-institute.org/accounts/dair" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>dair</span></a></span> </p><p>They've already published over 100 videos. If these haven't federated to your server yet, you can browse them all at <a href="https://peertube.dair-institute.org/a/dair/videos" rel="nofollow" class="ellipsis" title="peertube.dair-institute.org/a/dair/videos"><span class="invisible">https://</span><span class="ellipsis">peertube.dair-institute.org/a/</span><span class="invisible">dair/videos</span></a></p><p>You can also follow their Mastodon account at <span class="h-card"><a href="https://dair-community.social/@DAIR" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>[email protected]</span></a></span> </p><p><a href="/tags/featuredpeertube/" rel="tag">#FeaturedPeerTube</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/artificialintelligence/" rel="tag">#ArtificialIntelligence</a> <a href="/tags/peertube/" rel="tag">#PeerTube</a></p>


