I guess we shouldn't be surprised, but no way:<br><br>AAAI Launches AI-Powered Peer Review Assessment System<br><br><a href="https://aaai.org/aaai-launches-ai-powered-peer-review-assessment-system/" rel="nofollow" class="ellipsis" title="aaai.org/aaai-launches-ai-powered-peer-review-assessment-system/"><span class="invisible">https://</span><span class="ellipsis">aaai.org/aaai-launches-ai-powe</span><span class="invisible">red-peer-review-assessment-system/</span></a><br><br>No.<br><br>Speaking as someone who has co-organized an AAAI symposium and among other things did a bunch of editorial work.<br><br><a href="/tags/noai/" rel="tag">#NoAI</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/aioutofscience/" rel="tag">#aIOutOfScience</a> <a href="/tags/science/" rel="tag">#science</a> <a href="/tags/computerscience/" rel="tag">#ComputerScience</a> <a href="/tags/peerreview/" rel="tag">#PeerReview</a><br>
llms
<p>"When generative AI and large language models (LLMs) work, or appear to work, using them truly feels like magic. They can be very empowering, especially when used as an accessibility tool.</p><p>But they are also an unreliable technology built on exploitation of workers, particularly from third world countries, and our environment. Its frequent use may lead to cognitive decline and a loss of skills, and can actually decrease productivity."</p><p><a href="https://stefanbohacek.com/blog/on-generative-ai/" rel="nofollow" class="ellipsis" title="stefanbohacek.com/blog/on-generative-ai/"><span class="invisible">https://</span><span class="ellipsis">stefanbohacek.com/blog/on-gene</span><span class="invisible">rative-ai/</span></a></p><p><a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/ai/" rel="tag">#AI</a></p>
Edited 92d ago
<p>Recently, I spent a lot of time reading & writing about LLM benchmark construct validity for a forthcoming article. I also interviewed LLM researchers in academia & industry. The piece is more descriptive than interpretive, but if I’d had the freedom to take it where I wanted it to go, I would’ve addressed the possibility that mental capabilities (like those that benchmarks test for) are never completely innate; they’re always a function of the tests we use to measure them ...</p><p>(1/2)</p><p><a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/ai/" rel="tag">#AI</a></p>
<p>"This is just such a low tech, simple intervention, and can make people feel significantly less lonely."</p><p><a href="https://www.404media.co/chatgpt-loneliness-study-college-students-random-strangers-texting/" rel="nofollow" class="ellipsis" title="www.404media.co/chatgpt-loneliness-study-college-students-random-strangers-texting/"><span class="invisible">https://</span><span class="ellipsis">www.404media.co/chatgpt-loneli</span><span class="invisible">ness-study-college-students-random-strangers-texting/</span></a></p><p><a href="/tags/study/" rel="tag">#study</a> <a href="/tags/mentalhealth/" rel="tag">#MentalHealth</a> <a href="/tags/loneliness/" rel="tag">#loneliness</a> <a href="/tags/chatbots/" rel="tag">#ChatBots</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a></p>
<p>> For example, Google reduced our headline “I used the ‘cheat on everything’ AI tool and it didn’t help me cheat on anything” to just five words: “‘Cheat on everything’ AI tool.” It almost sounds like we’re endorsing a product we do not recommend at all.</p><p><a href="https://www.theverge.com/tech/896490/google-replace-news-headlines-in-search-canary-coal-mine-experiment" rel="nofollow" class="ellipsis" title="www.theverge.com/tech/896490/google-replace-news-headlines-in-search-canary-coal-mine-experiment"><span class="invisible">https://</span><span class="ellipsis">www.theverge.com/tech/896490/g</span><span class="invisible">oogle-replace-news-headlines-in-search-canary-coal-mine-experiment</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/google/" rel="tag">#google</a> <a href="/tags/search/" rel="tag">#search</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/enshittification/" rel="tag">#enshittification</a></p>
Edited 17d ago
I've been playing around with this set of ideas and questions:<br><br>An image of a cat is not a cat, no matter how many pixels it has. A video of a cat is not a cat, no matter the framerate. An interactive 3-d model of a cat is not a cat, no matter the number of voxels or quality of dynamic lighting and so on. In every case, the computer you're using to view the artifact also gives you the ability to dispel the illusion. You can zoom a picture and inspect individual pixels, pause a video and step through individual frames, or distort the 3-d mesh of the model and otherwise modify or view its vertices and surfaces, things you can't do to cats even by analogy. As nice or high-fidelity as the rendering may be, it's still a rendering, and you can handily confirm that if you're inclined to.<br><br>These facts are not specific to images, videos, or 3-d models of cats. The are necessary features of digital computers. Even theoretically. The computable real numbers form a countable subset of the uncountably infinite set of real numbers that, for now at least, physics tells us our physical world embeds in. Georg Cantor showed us there's an infinite difference between the two; and Alan Turing showed us that it must be this way. In fact it's a bit worse than this, because (most) physics deals in continua, and the set of real numbers, big as it is, fails to have a few properties continua are taken to have. C.S. Peirce said that continua contain such multitudes of points smashed into so little space that the points fuse together, becoming inseparable from one another (by contrast we can speak of individual points within the set of real numbers). Time and space are both continua in this way.<br><br>Nothing we can represent in a computer, even in a high-fidelity simulation, is like this. Temporally, computers have a definite cha-chunk to them: that's why clock speeds of CPUs are reported. As rapidly as these oscillations happen relative to our day-to-day experience, they are still cha-chunk cha-chunk cha-chunk discrete turns of a ratchet. There's space in between the clicks that we sometimes experience as hardware bugs, hacks, errors: things with negative valence that we strive to eliminate or ignore, but can never fully. Likewise, even the highest-resolution picture still has pixels. You can zoom in and isolate them if you want, turning the most photorealistic image into a Lite Brite. There's space between the pixels too, which you can see if you take a magnifying glass to your computer monitor, even the retina displays, or if you look at the data within a PNG.<br><br>Images have glitches (e.g., the aliasing around hard edges old JPEGs had). Videos have glitches (e.g., those green flashes or blurring when keyframes are lost). Meshes have glitches (e.g., when they haven't been carefully topologized and applied textures crunch and distort in corners). 3-d interactive simulations have <a href="https://www.youtube.com/channel/UCto7D1L-MiRoOziCXK9uT5Q" rel="nofollow">unending glitches</a>. The glitches manifest differently, but they're always there, or lurking. <a href="https://www.dullien.net/thomas/weird-machines-exploitability.pdf" rel="nofollow">They are reminders</a>.<br><br>With all that said: why would anyone believe generative AI could ever be intelligent? The only instances of intelligence we know inhabit the infinite continua of the physical world with its smoothly varying continuum of time (so science tells us anyway). Wouldn't it be more to the point to call it an intelligence simulation, and to mentally maintain the space between it and "actual" intelligence, whatever that is, analogous to how we maintain the mental space between a live cat and a cat video?<br><br>This is not the say there's something essential about "intelligence", but rather that there are unanswered questions here that seem important. It doesn't seem wise to assume they've been answered before we're even done figuring out how to formulate them well.<br><br><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llms/" rel="tag">#LLMs</a><br>
A potentially interesting question: how much would the appearance of sentience or intelligence that LLMs can generate for some users explode if they were forced to have deterministic output?<br><br>In principle you could add a single "freeze the random seed" toggle to any of the major chatbots, and with that setting toggled on they would always return precisely the same output for a given input. Organisms and by extension humans cannot behave like this---no matter how stereotyped an organism's response may seem, it always differs, in however small a way, from a previous response---and the LLM's illusion should immediately be obvious by contrast. But, perhaps more interestingly for the folks who do think LLMs exhibit some form of sentience or intelligence: are we really meant to believe that a random number generator is the source of sentience or intelligence? You could hook up a random number generator to a machine that is otherwise deterministic and clearly not sentient or intelligent, and it suddenly becomes so? How do you explain that?<br><br><a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/genai/" rel="tag">#GenAI</a> <a href="/tags/generativeai/" rel="tag">#GenerativeAI</a> <a href="/tags/llms/" rel="tag">#LLMs</a><br>
Edited 27d ago
<p>"These multinationals are coming to rule and dominate here. It’s a very unfortunate supply chain, and my call today as data labelers is to build up on this—as we are fighting for labor rights, we are also fighting for the environment […] we are fighting big companies."</p><p><a href="https://www.404media.co/ai-is-african-intelligence-the-workers-who-train-ai-are-fighting-back/" rel="nofollow" class="ellipsis" title="www.404media.co/ai-is-african-intelligence-the-workers-who-train-ai-are-fighting-back/"><span class="invisible">https://</span><span class="ellipsis">www.404media.co/ai-is-african-</span><span class="invisible">intelligence-the-workers-who-train-ai-are-fighting-back/</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/bigtech/" rel="tag">#BigTech</a> <a href="/tags/labor/" rel="tag">#labor</a> <a href="/tags/workersrights/" rel="tag">#WorkersRights</a></p>
<p>"These companies have primarily made their chatbots “smarter” not by writing niftier code but by making them bigger: ramming more data through more powerful computer chips that use more electricity."</p><p><a href="https://www.theatlantic.com/magazine/2026/04/ai-data-centers-energy-demands/686064/?gift=IaTMqerMHTZL-ib0ALx6bWD9eABk8EFIwj_d2aNSQBU&utm_source=copy-link&utm_medium=social&utm_campaign=share" rel="nofollow" class="ellipsis" title="www.theatlantic.com/magazine/2026/04/ai-data-centers-energy-demands/686064/?gift=IaTMqerMHTZL-ib0ALx6bWD9eABk8EFIwj_d2aNSQBU&utm_source=copy-link&utm_medium=social&utm_campaign=share"><span class="invisible">https://</span><span class="ellipsis">www.theatlantic.com/magazine/2</span><span class="invisible">026/04/ai-data-centers-energy-demands/686064/?gift=IaTMqerMHTZL-ib0ALx6bWD9eABk8EFIwj_d2aNSQBU&utm_source=copy-link&utm_medium=social&utm_campaign=share</span></a></p><p><a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/environment/" rel="tag">#environment</a></p>
<p>"OpenAI’s own data show that use of ChatGPT was pretty evenly split between work and personal cases in 2024, but by 2025, 73 percent of conversations with ChatGPT were personal, not for work."</p><p><a href="https://www.theatlantic.com/family/2026/03/ai-friendship-chatbot/686345/?gift=IaTMqerMHTZL-ib0ALx6bVCqQpREXvmI4_jZ7t47BIM&utm_source=copy-link&utm_medium=social&utm_campaign=share" rel="nofollow" class="ellipsis" title="www.theatlantic.com/family/2026/03/ai-friendship-chatbot/686345/?gift=IaTMqerMHTZL-ib0ALx6bVCqQpREXvmI4_jZ7t47BIM&utm_source=copy-link&utm_medium=social&utm_campaign=share"><span class="invisible">https://</span><span class="ellipsis">www.theatlantic.com/family/202</span><span class="invisible">6/03/ai-friendship-chatbot/686345/?gift=IaTMqerMHTZL-ib0ALx6bVCqQpREXvmI4_jZ7t47BIM&utm_source=copy-link&utm_medium=social&utm_campaign=share</span></a></p><p><a href="/tags/article/" rel="tag">#article</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/psychology/" rel="tag">#psychology</a></p>
<p>"As part of this, we are reducing unnecessary Copilot entry points, starting with apps like Snipping Tool, Photos, Widgets and Notepad."</p><p>Nice to see that all that pushback has worked!</p><p><a href="https://blogs.windows.com/windows-insider/2026/03/20/our-commitment-to-windows-quality/" rel="nofollow" class="ellipsis" title="blogs.windows.com/windows-insider/2026/03/20/our-commitment-to-windows-quality/"><span class="invisible">https://</span><span class="ellipsis">blogs.windows.com/windows-insi</span><span class="invisible">der/2026/03/20/our-commitment-to-windows-quality/</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/microsoft/" rel="tag">#microsoft</a> <a href="/tags/windows/" rel="tag">#windows</a> <a href="/tags/enshittification/" rel="tag">#enshittification</a></p>
Edited 15d ago
<p><a href="/tags/lispygopherclimate/" rel="tag">#lispyGopherClimate</a> <a href="/tags/live/" rel="tag">#live</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/podcast/" rel="tag">#podcast</a> (?!) <a href="https://toobnix.org/w/jQkCWCeFNRL9Utcr2GWurM" rel="nofollow" class="ellipsis" title="toobnix.org/w/jQkCWCeFNRL9Utcr2GWurM"><span class="invisible">https://</span><span class="ellipsis">toobnix.org/w/jQkCWCeFNRL9Utcr</span><span class="invisible">2GWurM</span></a><br>Chat live in <a href="/tags/lambdamoo/" rel="tag">#lambdaMOO</a> as always <a href="https://lambda.moo.mud.org/" rel="nofollow"><span class="invisible">https://</span>lambda.moo.mud.org/</a><br>(@join screwtape<br>"hey<br>)</p><p>- I release NZ government secrets about <a href="/tags/ai/" rel="tag">#AI</a> and <a href="/tags/llms/" rel="tag">#LLMs</a> senior managers keep sending me</p><p>- I start /actually/ multimooing my own personal moos LambdaMOO</p><p>- Otherwise, my <a href="/tags/commonlisp/" rel="tag">#commonLisp</a> brain is entirely inside my <a href="/tags/dl/" rel="tag">#DL</a> <a href="/tags/deeplearning/" rel="tag">#DeepLearning</a> <a href="/tags/roc/" rel="tag">#roc</a> <a href="/tags/statistics/" rel="tag">#statistics</a> original formulation <a href="https://lispy-gopher-show.itch.io/dl-roc-lisp" rel="nofollow" class="ellipsis" title="lispy-gopher-show.itch.io/dl-roc-lisp"><span class="invisible">https://</span><span class="ellipsis">lispy-gopher-show.itch.io/dl-r</span><span class="invisible">oc-lisp</span></a></p><p><span class="h-card"><a href="https://climatejustice.social/@kentpitman" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>kentpitman</span></a></span> featuring.</p>
Edited 19d ago
<p>"While much of the economic value generated by AI remains concentrated in technological centres such as Silicon Valley, many of its environmental and social costs are in these territories."</p><p><a href="https://restofworld.org/2026/ai-pushback-chile-mexico-kenya-philippines/" rel="nofollow" class="ellipsis" title="restofworld.org/2026/ai-pushback-chile-mexico-kenya-philippines/"><span class="invisible">https://</span><span class="ellipsis">restofworld.org/2026/ai-pushba</span><span class="invisible">ck-chile-mexico-kenya-philippines/</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/datacenters/" rel="tag">#DataCenters</a> <a href="/tags/digitalcolonialism/" rel="tag">#DigitalColonialism</a> <a href="/tags/bigtech/" rel="tag">#BigTech</a></p>
<p>Boost plz!</p><p>Looking for critical scholarship on the use of "AI" by library/archive workers. University libraries in particular, but adjacent and tangentially-relevant-at-best stuff is welcome too. Any format is fine: books, papers, blogposts, whatever. If it's good, gimme all you've got!</p><p>Looks like we're gonna have a department-wide conversation about people using LLMs, and it's being framed as "we're all using it, but we're not talking about it, so let's make sure we're all on the same page about using it responsibly" ... I'll of course be pushing the "there's basically no way to use it responsibly" position, and I'd like to arm myself and others with some critical analyses of issues related to its use in library/archive spaces.</p><p><a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/ai/" rel="tag">#ai</a> <a href="/tags/libraries/" rel="tag">#libraries</a> <a href="/tags/archives/" rel="tag">#archives</a></p>
Edited 11d ago
<p>I just finished a short introduction post about <span class="h-card"><a href="https://come-from.mad-scientist.club/@algernon" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>algernon</span></a></span>'s project iocaine: the deadliest poison known to AI <img src="https://neodb.social/media/emoji/social.tchncs.de/blobcatsunglasses.png" class="emoji" alt=":blobcatsunglasses:" title=":blobcatsunglasses:"> (not man <img src="https://neodb.social/media/emoji/social.tchncs.de/blobcatgiggle.png" class="emoji" alt=":blobcatgiggle:" title=":blobcatgiggle:">)</p><p>Within the post I explain what iocaine is, how it's related to AI and LLMs and of course why I use it to fight AI/LLM companies and to poison their crawlers and training sets.</p><p>Using, configuring and watching iocaine was also a way for me to shifting my dystopic/pessimistic thoughts <img src="https://neodb.social/media/emoji/social.tchncs.de/blobcat_thisisfine.png" class="emoji" alt=":blobcat_thisisfine:" title=":blobcat_thisisfine:"> into fun and joy <img src="https://neodb.social/media/emoji/social.tchncs.de/ablobcatattention.png" class="emoji" alt=":ablobcatattention:" title=":ablobcatattention:"></p><p>And yes, I hate AI and LLMs and yes, I'm really fine by becoming an obsolete developer by not using it <img src="https://neodb.social/media/emoji/social.tchncs.de/ablobcatheart.png" class="emoji" alt=":ablobcatheart:" title=":ablobcatheart:"></p><p><a href="https://lukasrotermund.de/posts/fighting-ai-and-llms-with-iocaine/" rel="nofollow" class="ellipsis" title="lukasrotermund.de/posts/fighting-ai-and-llms-with-iocaine/"><span class="invisible">https://</span><span class="ellipsis">lukasrotermund.de/posts/fighti</span><span class="invisible">ng-ai-and-llms-with-iocaine/</span></a></p><p><a href="/tags/iocaine/" rel="tag">#iocaine</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/fightai/" rel="tag">#FightAI</a> <a href="/tags/selfhosting/" rel="tag">#selfhosting</a></p>
<p>"AI models that lie and cheat appear to be growing in number with reports of deceptive scheming surging in the last six months, a study into the technology has found."</p><p>Grok, is this true? You must tell me if this is not true.</p><p><a href="https://www.theguardian.com/technology/2026/mar/27/number-of-ai-chatbots-ignoring-human-instructions-increasing-study-says" rel="nofollow" class="ellipsis" title="www.theguardian.com/technology/2026/mar/27/number-of-ai-chatbots-ignoring-human-instructions-increasing-study-says"><span class="invisible">https://</span><span class="ellipsis">www.theguardian.com/technology</span><span class="invisible">/2026/mar/27/number-of-ai-chatbots-ignoring-human-instructions-increasing-study-says</span></a></p><p>via <a href="https://m.ai6yr.org/@ai6yr/116302171997057021" rel="nofollow" class="ellipsis" title="m.ai6yr.org/@ai6yr/116302171997057021"><span class="invisible">https://</span><span class="ellipsis">m.ai6yr.org/@ai6yr/11630217199</span><span class="invisible">7057021</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/enshittification/" rel="tag">#enshittification</a></p>
<p>Catching up with some of the news coming out of the Atmosphere conference.</p><p>"With Attie, anyone will be able to build their own custom feed just by typing in commands in natural language, the same as if they’re chatting with any other AI chatbot."</p><p>I'm guessing NFT profile pictures are next?</p><p><a href="https://techcrunch.com/2026/03/28/bluesky-leans-into-ai-with-attie-an-app-for-building-custom-feeds/" rel="nofollow" class="ellipsis" title="techcrunch.com/2026/03/28/bluesky-leans-into-ai-with-attie-an-app-for-building-custom-feeds/"><span class="invisible">https://</span><span class="ellipsis">techcrunch.com/2026/03/28/blue</span><span class="invisible">sky-leans-into-ai-with-attie-an-app-for-building-custom-feeds/</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/atmosphere/" rel="tag">#atmosphere</a> <a href="/tags/atproto/" rel="tag">#ATProto</a> <a href="/tags/bluesky/" rel="tag">#bluesky</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a></p>
<p>"Of the nearly 1,400 Americans surveyed, more than three-quarters said they don’t trust AI — 76% say they trust it rarely or only sometimes, compared to just 21% who trust it most or almost all of the time."</p><p><a href="https://techcrunch.com/2026/03/30/ai-trust-adoption-poll-more-americans-adopt-tools-fewer-say-they-can-trust-the-results/" rel="nofollow" class="ellipsis" title="techcrunch.com/2026/03/30/ai-trust-adoption-poll-more-americans-adopt-tools-fewer-say-they-can-trust-the-results/"><span class="invisible">https://</span><span class="ellipsis">techcrunch.com/2026/03/30/ai-t</span><span class="invisible">rust-adoption-poll-more-americans-adopt-tools-fewer-say-they-can-trust-the-results/</span></a></p><p><a href="/tags/news/" rel="tag">#news</a> <a href="/tags/technology/" rel="tag">#technology</a> <a href="/tags/technews/" rel="tag">#TechNews</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/enshittification/" rel="tag">#enshittification</a></p>
<p>一些 <a href="/tags/阴谋论/" rel="tag">#阴谋论</a> 在中文互联网上流传,认为一些失踪人口的意识被上传,以用于开发豆包等大语言模型。</p><p><a href="https://www.zhihu.com/question/2022623696783774161" rel="nofollow" class="ellipsis" title="www.zhihu.com/question/2022623696783774161"><span class="invisible">https://</span><span class="ellipsis">www.zhihu.com/question/2022623</span><span class="invisible">696783774161</span></a><br><a href="https://xcancel.com/YesterdayOrange/status/2038891964415037591" rel="nofollow" class="ellipsis" title="xcancel.com/YesterdayOrange/status/2038891964415037591"><span class="invisible">https://</span><span class="ellipsis">xcancel.com/YesterdayOrange/st</span><span class="invisible">atus/2038891964415037591</span></a></p><p><a href="/tags/conspiracy/" rel="tag">#conspiracy</a> <a href="/tags/internetmysteries/" rel="tag">#internetmysteries</a> <a href="/tags/missingperson/" rel="tag">#missingperson</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/互联网谜团/" rel="tag">#互联网谜团</a> <a href="/tags/失踪人口/" rel="tag">#失踪人口</a> <a href="/tags/人工智能/" rel="tag">#人工智能</a> <a href="/tags/大语言模型/" rel="tag">#大语言模型</a> <a href="/tags/llm/" rel="tag">#llm</a> <a href="/tags/llms/" rel="tag">#llms</a></p>
Edited 5d ago
<p>DAIR is a research institute that is highly sceptical about AI hype and the big tech companies behind it. You can follow their excellent video account at:</p><p>➡️ <span class="h-card"><a href="https://peertube.dair-institute.org/accounts/dair" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>dair</span></a></span> </p><p>They've already published over 100 videos. If these haven't federated to your server yet, you can browse them all at <a href="https://peertube.dair-institute.org/a/dair/videos" rel="nofollow" class="ellipsis" title="peertube.dair-institute.org/a/dair/videos"><span class="invisible">https://</span><span class="ellipsis">peertube.dair-institute.org/a/</span><span class="invisible">dair/videos</span></a></p><p>You can also follow their Mastodon account at <span class="h-card"><a href="https://dair-community.social/@DAIR" class="u-url mention" rel="nofollow noopener noreferrer" target="_blank">@<span>[email protected]</span></a></span> </p><p><a href="/tags/featuredpeertube/" rel="tag">#FeaturedPeerTube</a> <a href="/tags/ai/" rel="tag">#AI</a> <a href="/tags/llm/" rel="tag">#LLM</a> <a href="/tags/llms/" rel="tag">#LLMs</a> <a href="/tags/artificialintelligence/" rel="tag">#ArtificialIntelligence</a> <a href="/tags/peertube/" rel="tag">#PeerTube</a></p>
