AI Chatbots Remain Overconfident — Even When They’re Wrong: Large Language Models appear to be unaware of their own mistakes, prompting concerns about common uses for AI chatbots.
AI Chatbots Remain Overconfident — Even When They’re Wrong - Dietrich College of Humanities and Social Sciences - Carnegie Mellon University
Large Language Models appear to be unaware of their own mistakes, prompting concerns about common uses for AI chatbots.www.cmu.edu
like this
Tim Chambers reshared this.
AI Chatbots Remain Overconfident — Even When They’re Wrong: Large Language Models appear to be unaware of their own mistakes, prompting concerns about common uses for AI chatbots.
AI Chatbots Remain Overconfident — Even When They’re Wrong - Dietrich College of Humanities and Social Sciences - Carnegie Mellon University
Large Language Models appear to be unaware of their own mistakes, prompting concerns about common uses for AI chatbots.www.cmu.edu
‘Japanese-first’ Sanseito party goes into election leveraging unease about foreigners
How Japan’s hard-right populists are profiting from anti-foreign sentiment and a cost of living crunch
Nationalists win over disaffected first-time voters with a call for a return to family values and curbs on immigrationGavin Blair (The Guardian)
Some thoughts on Surf, Flipboard's fediverse app
I've got access to the beta of the Surf app. Some thoughts:
some stuff I really liked:
- rss works (though no custom URLs yet, just what they already scraped)
- you get lemmy, mastodon, bluesky, threads all together
- you can make your own feeds and check what other people made (like a custom timeline, or topic-specific like “NBA”, “woodworking”, “retro gaming stuff”)
- has different modes: you can switch between videos, articles, podcasts depending on the feed
but also...
- can’t add your own RSS feeds (huge miss)
- some feeds break and show no posts even when they’re active (ok, it's still a beta)
- YouTube videos have ads (not into that—I support creators through patreon, affiliate links, whatever. not ads)
- feeds you create are public by default unless you manually change it
- not open source. built on open protocols, sure. but the app is locked up. (HUGE MISS)
all that said, I really believe: better feeds = better experience = better shot at the fediverse going mainstream.
anyone else tried it?
do you know anyone building an open source version of this? is that even realistic?
I’d love to hear what do you think 😀
Flow control? China starts mega-dam project on Brahmaputra in Tibet; how will it impact India - Times of India
Flow control? China starts mega-dam project on Brahmaputra in Tibet; how will it impact India
China News: China has commenced construction of a major dam on the Brahmaputra River in Tibet, near the Indian border, with Premier Li Qiang present at the groundTOI World Desk (The Times Of India)
London: Over 50 arrests in Parliament Square amid pro-Palestine Action protest
More than 50 arrests in Parliament Square as pro-Palestine Action protests held across UK
Dozens of protesters assembled in central London on Saturday afternoonSami Quadri (Evening Standard)
Reddit users in the UK must now upload selfies to access NSFW subreddits
Reddit is introducing age verification for UK users
The change is due to new age verification laws in the UK.Amanda Yeo (Mashable)
They can force custom proprietary spying software on your devices.
- That would block Linux from their borders, which means goodbye Steam Deck in the UK among other things.
Migrants sent to El Salvador's CECOT returned to Venezuela in prisoner swap, 10 Americans freed: Officials
Migrants sent to El Salvador's CECOT returned to Venezuela in prisoner swap, 10 Americans freed: Officials
Over 250 prisoners were released from CECOT, Venezuela's government said.Laura Romero (ABC News)
Channel.org open beta
Seems to be a way of making Bluesky style feeds with Mastodon-style services, well that's what I gather from reading the FAQ. They don't actually explain what this is anywhere.
Channel is basically a white label instance of PatchWork, which is a Mastodon fork with custom feeds and community curation tools.
The main intent behind the project is to help existing communities and organizations get onto the Fediverse, and have some curation capabilities. Ideally, it can be used to get a large amount of people and accounts onto the network with minimal friction.
GitHub - patchwork-hub/patchwork-web: Your self-hosted, globally interconnected microblogging community
Your self-hosted, globally interconnected microblogging community - patchwork-hub/patchwork-webGitHub
An American father who moved to Russia to avoid LGBTQ+ “indoctrination” is being sent to the front line against Ukraine despite being assured he would serve in a non-combat role.
Anti-Woke Dad Who Fled With Family to Russia Sent to War Zone
Derek Huffman, 46, joined the military with hopes of becoming a Russian citizen. His wife said he was duped into a combat role.Josh Fiallo (The Daily Beast)
We really need to stop abandoning existing foss projects and thinking a whole new thing needs to be invented. Free and open-source software is not a product, it doesn't abide by the same rules and relationships that proprietary tech does.
It's more organic. It's also a commons that we can continue to draw on, and reshape. If I recall correctly, there were something like three different vector graphic editors from the same codebase before Inkscape managed to be the one that gained traction.
Matrix isn't perfect, but abandoning it just to reinvent it all over again just because some people really need a thing that works like Discord, even though Discord is absolute hot garbage; is just going to re-create all the same problems. Matrix today is better than it was two years ago. And Matrix in a year will be better from now.
Honestly, setting up things using Docker Compose is generally a question of copying and pasting and editing the file locations.
The moment you need SSL and/or a reverse proxy it becomes a bit more complex, but once you set up a reverse proxy once you can generally expand that to your other applications.
Something like a Synology nas makes it very easy and to some extend even the Truenas apps are kinda easy.
Knowledge manipulation on Russia's Wikipedia fork; Marxist critique of Wikidata license; call to analyze power relations of Wikipedia
Disposable E-Cigarettes More Toxic Than Traditional Cigarettes: High Levels of Lead, Other Hazardous Metals Found in E-Cigarettes Popular with Teens
They may look like travel shampoo bottles and smell like bubblegum, but after a few hundred puffs, some disposable, electronic cigarettes and vape pods release higher amounts of toxic metals than older e-cigarettes and traditional cigarettes, according to a study from the University of California, Davis. For example, one of the disposable e-cigarettes studied released more lead during a day’s use than nearly 20 packs of traditional cigarettes.
Disposable E-Cigarettes More Toxic Than Traditional Cigarettes
They may look like travel shampoo bottles and smell like bubblegum, but after a few hundred puffs, some disposable, electronic cigarettes and vape pods release higher amounts of toxic metals than older e-cigarettes and traditional cigarettes, accordi…UC Davis
Xinjiang’s Organ Transplant Expansion Sparks Alarm Over Uyghur Forced Organ Harvesting
cross-posted from: sh.itjust.works/post/42460866
Xinjiang’s official organ donation rate is shockingly low. So why is China planning to open six new organ transplant facilities in the region"The expansion suggests that the Chinese authorities are expecting to increase the numbers of transplants performed in Xinjiang. However, this is puzzling as there is no reason why the demand for transplants should suddenly go up in Xinjiang,” Rogers explained. “From what we know about alleged voluntary donations, the rates are quite low in Xinjiang. So the question is, why are these facilities planned?”
Rogers noted one chilling possibility: that “murdered prisoners of conscience (i.e., Uyghurs held in detention camps)” could be a source of transplanted organs.
This suggestion becomes even more concerning when considering the extensive surveillance and repression that Uyghurs face in the region. Detainees in the many internment camps in Xinjiang have reported being subjected to forced blood tests, ultrasounds, and organ-focused medical scans. These procedures align with organ compatibility testing, raising fears that Uyghurs are being prepped for organ harvesting while in detention.
David Matas, an international human rights lawyer who has investigated forced organ harvesting in China, questioned the very possibility of voluntary organ donation in Xinjiang. “The concept of informed, voluntary consent is meaningless in Xinjiang’s carceral environment,” Matas said. “Given the systemic repression, any claim that donations are voluntary should be treated with the utmost skepticism.”
The new transplant facilities will be distributed across Urumqi and other regions of northern, southern, and eastern Xinjiang. Experts argue that the sheer scale of this expansion is disproportionate to Xinjiang’s voluntary donation rate and overall capacity, suggesting that the Chinese authorities may be relying on unethical methods to source organs.
like this
The wikipedia article for the universal peace federation redirects to the unification church article.
Shinzo Abe found out how bad the moonies are.
Keep spreading that "new cold war" propaganda.
Nobody is defending the Moonies, especially not this current affairs publication owned by a Japanese media corporation. Here's plenty of examples of them calling out the Unification Church:
thediplomat.com/tag/unificatio…
Anybody can be nominated to be an ambassador for peace, it's also associated with the UN.
upf.org/core-program/ambassado…
Launched in 2001, Ambassadors for Peace is the largest and most diverse network of peace leaders. As of 2020, there are more than 100,000 Ambassadors for Peace from 160 countries who come from all walks of life representing many races, religions, nationalities, and cultures
Literally she has no other ties to the Moonies/unification church, and how about the human right lawyer she directly quotes.
Or the bioethicist and part of the coalition to End Transplant Abuses in China (ETAC)? All just cold war propaganda?
Results 445 included studies reported on outcomes of 85 477 transplants. 412 (92.5%) failed to report whether or not organs were sourced from executed prisoners; and 439 (99%) failed to report that organ sources gave consent for transplantation. In contrast, 324 (73%) reported approval from an IRB. Of the papers claiming that no prisoners’ organs were involved in the transplants, 19 of them involved 2688 transplants that took place prior to 2010, when there was no volunteer donor programme in China.
Anyway, keep spreading that there is no genocide propaganda.
washingtonpost.com/politics/20…
Two months after the Trump administration all but shut down its foreign news services in Asia, China is gaining significant ground in the information war, building toward a regional propaganda monopoly, including in areas where U.S.-backed outlets once reported on Beijing’s harsh treatment of ethnic minorities.The U.S. decision to shut down much of RFA’s shortwave broadcasting in Asia is one of several cases where the Trump administration — which views China as America’s biggest rival — has yielded the adversary a strategic advantage.
Allentown grandfather’s family was told he died in ICE custody. Then they learned he’s alive — in a hospital in Guatemala, they say
The Eighth Amendment to the United States Constitution states: “Excessive bail shall not be required, nor excessive fines imposed, nor cruel and unusual punishments inflicted.”
Unconstitutional actions ordered by the POTUS. Are we ready to impeach yet?
https://www.mcall.com/2025/07/18/luis-leon-allentown-grandfather-ice-guatemala/
The man was granted asylum! He’s 82 god damn years old!
I would love to see a popular uprising where we string up the thugs that are snatching people off the streets. They don’t need unnecessary things like “lawyers” or “trials”. A can of gas and a match are pretty cheap. So is rope.
Anybody else not able to get on slrpnk.net?
Seems like slrpnk.net hasn't been working for most of the day today. Hasn't worked for me on mobile or desktop. Says 502 bad gateway when trying to access the website. Anybody else expierencing this issue?
Edit: it's back up. Thank you admins!
like this
thanks for passing that along. I could tell something was up by looking at this:
like this
In all seriousness, to the users and admins of slrpnk.net, you have my solidarity. Hope this all gets resolved soon.
Edit: oh shit welcome back!
UEA sekretigas la elekton de kongresurboj
Laŭ la kongresa regularo de UEA, la komitato estu regule informata kaj konsultata pri la elekto de kongresurboj. Laŭ la nova prezidanto de UEA, Fernando Maia, tio tamen ne eblas, ĉar la kandidata urbo ne sciu, ĉu ĝi estas la sola kandidato. Tial la regularo laŭ li devas esti ŝanĝita.
The Hype is the Product
The Hype is the Product
Large publicly traded tech companies seem to no longer consider their customers – that is, people and organizations who actually buy their products or pay for access to their services – their core focSongs on the Security of Networks
adhocfungus likes this.
introducing copyparty, the FOSS file server
- YouTube
Profitez des vidéos et de la musique que vous aimez, mettez en ligne des contenus originaux, et partagez-les avec vos amis, vos proches et le monde entier.www.youtube.com
thisisbutaname likes this.
OpenAI Is Giving Exactly the Same Copy-Pasted Response Every Time Time ChatGPT Is Linked to a Mental Health Crisis
OpenAI Is Giving Exactly the Same Copy-Pasted Response Every Time Time ChatGPT Is Linked to a Mental Health Crisis
As reports of ChatGPT sending its users down dangerous mental health spirals mount, OpenAI seemingly can only think of one thing to say.Frank Landymore (Futurism)
like this
This feels like a great time to recommend a song by a parody-hate band, S.O.D.:
Please understand that this band was formed by Scott Ian, of Anthrax, in the 80s. This was a time when you could mock hateful racists and people understood that it was a joke. I wouldn't support a band saying that now, because I'd consider the excuse that it was a joke to be a front for their actual beliefs, as we've seen with people who are "just asking questions."
Anthrax and Public Enemy teamed up on Bring Tha Noise because Anthrax liked rap. Aerosmith teamed up with Run DMC because their manager / producer / someone convinced them to. Anthrax was genuinely not about hate.
Bonus trivia: Scott Ian now plays with Mr Bungle. Just as S.O.D's titular song was called Speak English or Die, Mr Bungle now plays a song called Habla Español O Muere (Speak Spanish or Die). If you can't judge that the former was a parody by the evolution of the theme, I don't know what to tell you.
Edit: formatting and more info.
- YouTube
Profitez des vidéos et de la musique que vous aimez, mettez en ligne des contenus originaux, et partagez-les avec vos amis, vos proches et le monde entier.www.youtube.com
ah, dear old copy/paste.... It's funny that even OpenAI doesn't trust ChatGPT enough to give more personalized LLM-generated answers.
And this sounds exactly like the type of use case AI agents are supposedly so great at that they will replace all human workers (according to Altman at least). Any time now!
Nvidia plans to boost presence in Israel with multibillion-dollar tech campus in north
Nvidia is actively seeking land to build a massive multibillion-dollar tech campus in Israel’s north, which is expected to provide thousands of jobs in what promises to be a major expansion of the US chip giant’s operations in the country.
The computing juggernaut announced on Sunday that it had issued a so-called request for information (RFI) tender to locate a plot of land spanning 70 to 120 dunams (30 acres) with construction rights to build a campus of 80,000–180,000 square meters. Nvidia is interested in buying land with “high accessibility to main traffic arteries and public transportation” around Zichron Yaakov, Haifa, and the Jezreel Valley areas. The tech titan has hired real estate consulting firm Colliers for the search and has set July 23 as the deadline for submissions.
like this
Premio musicale aulla
Quanto costa il premio Lunezia? | Eco della Lunigiana
Il 25 luglio Piazza Gramsci tornerà a riempirsi di musica, Aulla è pronta ad ospitare una delle tappe della 30ª edizione del Premio Lunezia, con due nomi cheDiego Remaggi (Eco della Lunigiana)
Deadly airdrops and a trickle of trucks won't undo months of 'engineered starvation' in Gaza, Oxfam says
July 27, 2025 09:28 EDT
Oxfam has said the airdrops into #Gaza are wholly inadequate for the population’s needs and has called for the immediate opening of all crossings for full humanitarian access into the territory devastated by relentless #Israeli bombardments and a partial aid blockade.
Bushra Khalidi, Oxfam policy lead for the Occupied #Palestinian territory, said:
Deadly airdrops and a trickle of trucks won’t undo months of engineered starvation in Gaza.What’s needed is the immediate opening of all crossings for full, unhindered, and safe aid delivery across all of Gaza and a permanent ceasefire. Anything less risks being little more than a tactical gesture.
Middle East crisis live: Israeli military announces ‘tactical pause’ in parts of Gaza as pressure mounts over hunger
Military says it will halt activity in Muwasi, Deir al-Balah and Gaza City from 10am to 8pm local time every day until further noticeGuardian staff reporter (The Guardian)
like this
Deadly airdrops and a trickle of trucks won't undo months of 'engineered starvation' in Gaza, Oxfam says
cross-posted from: lemmy.ml/post/33751786
July 27, 2025 09:28 EDTOxfam has said the airdrops into #Gaza are wholly inadequate for the population’s needs and has called for the immediate opening of all crossings for full humanitarian access into the territory devastated by relentless #Israeli bombardments and a partial aid blockade.
Bushra Khalidi, Oxfam policy lead for the Occupied #Palestinian territory, said:
Deadly airdrops and a trickle of trucks won’t undo months of engineered starvation in Gaza.What’s needed is the immediate opening of all crossings for full, unhindered, and safe aid delivery across all of Gaza and a permanent ceasefire. Anything less risks being little more than a tactical gesture.
Updates for controlled mechanical ventilation system (double flux) without privileged admin
[Update in the comments]
Hello all!
I've got a controlled mechanical ventilation system (system D) at home from Zehnder (ComfoAir Q600). I've even got their controller box (the LAN-C) so I can use smart home stuff with it. It works perfect on home assistant, even when blocking the controller on the router level from the outside world. Maintenance wise, they try to force a contract on you, but it is easy peasy to maintain and repair so I'm not having no maintenance contract.
Comes the issue of software and updates. Some updates come with features. Sometimes, they are even mandatory so addons can work on them (ex: small heat pump for the intake needs a recent version for setup). For this, you have to use their app on your phone/tablet. The whole idea is that the install goes trough your phone (with checksum check through the app) to the EEROM on the local network to prevent bricking of the unit. Updates bring usually nice settings, and are sometimes mandatory for some add-ons (ex: heatpump for pre-heating or pre-cooling needs a recent update to be set up).
Here comes the really annoying part that makes me grump a lot: to update, of even for some diagnose option, you need to access a special level. Not the beginner mode. Not the expert mode. Not the installer mode that is unlocked with a simple pin code available in the owner's manual. No sweet child, you need to be a registered installer with Zehnder to access to get updates and real diagnostics. Officially, it is to prevent bricking the controller with an update by an user. But it is possible to give access to a licensed installer so they can update remotely and run diagnostics. So an issue with your internet, and there is no more safeguard to protect you from bricking stuff. Really, it is just to force a maintenance visit (200€ to exchange filters and clean a bit the exchanger and the inside with some soapy water). I don't like to bend over while I'm getting fucked without my consent, so you guess while this pisses me off. I called once to get an update (some companies ask you a hefty sum for that), but instead of getting an account they just updated it once exceptionally.
There is tho in the official documents for Germany, a test code publicly accessible, that allows you to access diagnostics and updates. But the updates there are only for German units. Pretty sure it is the same unit for the whole damn continent, but hey, let's pretend the units are different.
Comes my question: how do I trick the system into believing their update is not for the germans, but for somewhere else? Or even better, to give me access for updates for other areas? I know part is server side (account), but I'm willing to bet they don't really care about securing access to the updates once you have authenticated yourself (with the german test code). Tried lucky patcher, but didn't get lucky.
Any idea what I could try (even Lucky Patcher wise)?
Big hugs and kisses
If there's a German code that would work as you intended (if I got you right) but it doesn't for you, since you don't live in Germany, would it work to make the machine believe it is located in Germany?
They might have hardcoded a location into it, then you are out of luck. But maybe they determine it via the internet connection you use to update? So you could potentially have it connected to a VPN through your router, which fakes a German location? Probably too simple a solution.
[Update]So used my old rooted tablet to tweak around a bit with the app. Through lucky patcher (when logged in with the test account) I noticed that the downloads are done trough the root user of android. After that I used MiXplorer to get the data files on my pc. Quickly found the structure of the files. Couldn't trick the system to access my local files, but I managed to trick the system into updating as if it were a german system.
So if someone else happens to look and stumble upon this, this is how I got it to work. It works only from a rooted android device for now:
- Login with the german test account to access server downloads
- Connect to the cloud delivery system and download the update that you want
- Close the app, with a root file explorer (like MiXplorer + Shizuku) go into the root folder (use their FTP server with a root allowed user or whatever to transfer it more easily to the PC).
- Go to /data/data/com.zehndergroup.comfocontrol/files/products/1/R1.12.0-DE
-Open the meta.json file and change the german id of your unit to your unit. Ex: 471502013 to 471502023 for the UK id. Save it and send it back to where it came from. You could just update your unit, and it would keep the same serial number, same everything but would be under german ID. Easy for new updates but annoying to explain if you need to have a technician over and he is wondering why your unit has that ID. But then again, that is a minor detail and I'm not even sure the technician will be paid enough to care. Reverting to your national number should be the same process but with the update for your country.
What didn't work:
*Open the config bin file of your unit (so again, for the Q600 : config_R1.12.0_471502013_v1.bin ) with a hex editor. Look for the unit number that needs to be replaced (so here 471502013 needs to become 471502023). I only needed to replace 1 number (a 1 into 2) , so 31 became 32 in the hex file. Replace the country code with your local one in hex (So DE into UK). Save it and send it back from where it came from. This provokes an error after the 3rd block. Probably a checksum that isn't cooperating in another file
*Seperate API connection: the naming pattern o their website is obvious, but connection without their app is something else
- Firmware updates for the ventilation units are in folder "1", maybe that will change in the future
- The downloaded firmware update will be there under it's own folder (like R1.12.0) and sometimes there will be it's own ZIP
- National ID for your unit is on Zehnder's website but also under "basic mode" > "unit status" > "Article unit"
- The installers pin code changes from your countries to the German one, so it becomes 4210
PS: there are ati-bricking measures in place in the system. If an update fails, you can Erase the firmware and reupload it but you'll have to redo the post-install setup
ACE & MPA Continued to Scoop Up Pirate Domains in Bulk During Q2 2025
In Q2 2025, anti-piracy coalition Alliance for Creativity and Entertainment continued to 'seize' domains in bulk, adding to the world's largest collection of former pirate domains maintained by the MPA. While the archive contains countless unique and memorable domains, many with interesting and informative backstories, new additions illustrate typical responses to site blocking measures and very little else.
ACE & MPA Continued to Scoop Up Pirate Domains in Bulk During Q2 2025 * TorrentFreak
The Alliance for Creativity and Entertainment continues to 'seize' pirate domains, adding to huge collection maintained by the MPA.Andy Maxwell (TF Publishing)
like this
With all these domains they are continuing to seize and then need to continue to renew, at what point does a bean counter tell them it's unsustainable? I also wonder how many sites just don't get renewed because they slipped through the cracks like certs sometimes do. I guess if you're a domain register this seems like a crazy inflated sales bubble that at some point is going to pop, and hopefully they have saved some of that revenue to ride out that lull.
Edit - a word
ACE & MPA Continued to Scoop Up Pirate Domains in Bulk During Q2 2025
In Q2 2025, anti-piracy coalition Alliance for Creativity and Entertainment continued to 'seize' domains in bulk, adding to the world's largest collection of former pirate domains maintained by the MPA. While the archive contains countless unique and memorable domains, many with interesting and informative backstories, new additions illustrate typical responses to site blocking measures and very little else.
ACE & MPA Continued to Scoop Up Pirate Domains in Bulk During Q2 2025 * TorrentFreak
The Alliance for Creativity and Entertainment continues to 'seize' pirate domains, adding to huge collection maintained by the MPA.Andy Maxwell (TF Publishing)
In un mese sono morte più di 80 persone in montagna - Il Post
circa la metà delle persone recuperate si rifiuta di pagare «anche quando, di fatto, gli hai salvato la vita»
Che roba! Nemmeno se gli salvi la vita sono disposti a pagare i soccorsi
In un mese sono morte più di 80 persone in montagna
Lo ha detto il presidente del Soccorso alpino nazionale, segnalando situazioni al limite con chi viene salvato che si rifiuta di pagareIl Post
reshared this
La deputata democratica April McClain Delaney lancia l'allarme: i tagli apportati da Donald Trump a programmi come Medicaid, nonché a NPR e PBS, colpiranno le zone rurali americane come uno "tsunami"
In un'intervista rilasciata a Newsweek, la deputata April McClain Delaney ha lanciato l'allarme: i tagli apportati da Donald Trump a programmi come Medicaid, nonché a NPR e PBS, colpiranno le zone rurali americane come uno "tsunami".
Il distretto congressuale del Maryland di Delaney comprende alcune delle aree che potrebbero essere maggiormente colpite dalle politiche di Trump. Si estende dalla zona rurale occidentale dello stato, che secondo lei potrebbe subire il peso dei nuovi tagli alla rescissione, alla periferia di Washington, DC, dove risiedono i dipendenti federali che hanno perso il lavoro a causa dei licenziamenti di massa.
"Se si considerano tutti questi congelamenti dei finanziamenti per i dipendenti pubblici dei nostri parchi nazionali, ma anche per Medicaid, SNAP e poi si cominciano a considerare alcune delle altre rescissioni, ci si rende conto che si tratta semplicemente di uno tsunami che sta per colpire l'America rurale", ha affermato Delaney.
Exclusive: April McClain Delaney Warns Trump Cuts to Hit Rural America Like 'a Tsunami'
Rep. April McClain Delaney told Newsweek how cuts to programs like Medicaid, as well as NPR and PBS, will affect rural Americans.Andrew Stanton (Newsweek)
like this
How we Rooted Copilot
How we Rooted Copilot - Eye Research
Read how we explored the Python sandbox in Copilot and got root on the underlying containerVaisha Bernard (Eye Security)
like this
Sunday, July 27, 2025
Share
The Kyiv Independent [unofficial]
This newsletter is brought to you by Medical Bridges.
Medical Supplies for Ukraine’s Hospitals. Partnering for global health equity.
Russia’s war against Ukraine
Debris litters a sports complex after an overnight Russian bombardment on July 26, 2025 in Kharkiv, Ukraine. Officials say this five-hour bombardment in the Kyivskyi district included four Russian glide bombs, two ballistic missiles, and 15 Shahed drones. (Scott Peterson/Getty Images)
‘Resistance inside Russia is growing’ — Su-27UB jet set alight in Krasnodar Krai, Ukraine’s HUR claims. “Resistance to the Kremlin regime inside Russia is growing,” Ukraine’s military intelligence agency said.
After being battered by Ukraine, Russia hopes to ‘strengthen’ Black Sea Fleet. “In the coming years, the Black Sea sailors will be further strengthened — with the arrival of new frigates, corvettes, aviation, marine robotic complexes,” Nikolai Patrushev, the head of Russia’s Maritime Collegium, said.
Ukraine reports killing Russian colonel leading assaults in Kharkiv Oblast. According to operational data by Ukraine’s Khortytsia group of forces, Colonel Lebedev was leading assault operations in the Velykyi Burluk area of Kharkiv Oblast.
Your contribution helps keep the Kyiv Independent going. Become a member today.
Russian drone strike damages Regional Military Administration building in Sumy. Russian forces launched a drone strike on Sumy on July 26, damaging the building of the Sumy Regional Military Administration, regional governor Oleh Hryhorov reported.
Ukraine ‘thwarts Russian plan for Sumy Oblast,’ Zelensky says. Ukrainian forces have pushed back Russian troops in Sumy Oblast, disrupting Moscow’s attempts to expand its foothold in the region, President Volodymyr Zelensky said on July 26.
Anti-corruption
How effective were Ukraine’s anti-corruption agencies targeted by Zelensky, and who were they investigating? The National Anti-Corruption Bureau (NABU) and the Specialized Anti-Corruption Prosecutor’s Office (SAPO) have investigated top officials, including Zelensky’s allies, and have widely been seen as more effective than other law enforcement agencies. However, real progress has been hampered by Ukraine’s flawed judicial system.
Our readers’ questions about Ukraine’s anti-corruption bodies, answered. We offered members of the Kyiv Independent community to share their questions about a controversial bill that undermined Ukraine’s anti-corruption institutions and the street protests that followed it this week.
Read our exclusives
Meet Ukraine’s EuroMaidan protesters fighting again for democracy in wartime Kyiv
Twelve years after the EuroMaidan Revolution, thousands mobilized across Ukraine again, united by a different cause. Protests erupted on the evening of July 22, just hours after Ukraine’s parliament passed a bill widely seen as an assault on corruption reform.
Photo: Anastasia Verzun / The Kyiv Independent
Learn more
‘Stop fueling Russia’s aggression’ — US, China clash over Ukraine at United Nations
The U.S. urged China to stop enabling Russia’s war in Ukraine during a UN Security Council meeting, prompting a sharp rebuke from Beijing, which accused Washington of creating confrontation.
Photo: Wang Fan/China News Service/Getty Images
Learn more
Orban offers Ukraine ‘strategic cooperation,’ claims EU accession would ‘drag the war’ into Europe
Hungarian Prime Minister Viktor Orban proposed “strategic cooperation” with Ukraine instead of European Union integration, arguing that Kyiv’s accession would drag the war into the heart of Europe.
Photo: Attila Kisbenedek / AFP
Learn more
Ukraine’s F-16 have a new trick to avoid Russian ballistic missiles
Ukraine’s fleet of F-16 fighter jets have been given a badly-needed boost with the creation of new mobile maintenance and operations modules which will help them evade Russian ballistic missile strikes.
Photo: Come Back Alive Charity Foundation
Human cost of Russia’s war
Russian forces attack Sumy Oblast, injure 3 civilians. Russian forces launched an overnight attack on Ukraine’s northeastern Sumy Oblast on July 26, targeting civilian infrastructure and leaving three people injured.
General Staff: Russia has lost 1,048,330 troops in Ukraine since Feb. 24, 2022. The number includes 1,080 casualties Russian forces suffered just over the past day.
9 killed, 61 injured in Russian attacks on Ukraine over past day. Ukraine’s Air Force said Russia launched 208 drones and 27 missiles overnight, targeting cities and infrastructure in multiple regions.
International response
US Senator Blumenthal warns Zelensky over anti-graft law, backs protests as ‘democracy in action.’ U.S. Democratic Senator Richard Blumenthal co-authored a bipartisan bill that would impose 500% tariffs on countries buying Russian oil, gas, or uranium.
Pope Leo meets Russian Orthodox cleric to discuss Ukraine war. The Vatican confirmed that Pope Leo received Metropolitan Anthony, the senior Russian Orthodox Church cleric who chairs its department of external church relations, along with five other high-profile clerics, during a morning audience on July 26.
Lithuania to allocate $32 million toward joint purchase of Patriot missile systems for Ukraine. Lithuania plans to contribute up to 30 million euro ($32.5 million) toward the joint purchase of U.S.-made Patriot air defense systems for Ukraine, Lithuanian public broadcaster LRT reported on July 26.
Russia could be ready for ‘confrontation with Europe‘ by 2027, Polish prime minister says. “Russia will be ready for confrontation with Europe — and therefore with us — as early as 2027,” Polish Prime Minister Donald Tusk said.
In other news
Lukashenko resumes use of migrants to ‘exert political pressure’ on Europe, Ukraine’s intelligence says. Belarusian President Alexander Lukashenko is facilitating transit primarily to Poland, with some migrants arriving from Libya directly or through Russia — pointing to coordinated efforts between Minsk and Moscow.
Ukrainian drones strike major Russian military radio factory in Stavropol, SBU source says. Ukrainian drones struck the Signal radio plant in Russia’s Stavropol Krai overnight on July 26. The plant, located around 500 kilometers (311 miles) from Ukraine-controlled territory, manufactures electronic warfare equipment for front-line aircraft and is a major part of Russia’s military-industrial complex.
The Kyiv Independent delivers urgent, independent journalism from the ground, from breaking news to investigations into war crimes. Your support helps us keep telling the truth. Become a member today.
This newsletter is open for sponsorship. Boost your brand’s visibility by reaching thousands of engaged subscribers. Click here for more details.
Today’s Ukraine Daily was brought to you by Tim Zadorozhnyy, Olena Goncharova, Sonya Bandouil, Oleg Sukhov, Irynka Hromotska, Kateryna Hodunova, Yuliia Taradiuk, and Lucy Pakhnyuk.
If you’re enjoying this newsletter, consider joining our membership program. Start supporting independent journalism today.
Share
Ukrainian drones strike major Russian military radio factory in Stavropol, SBU source says
The plant manufactures countermeasure stations for front-line aircraft and is a major part of Russia's military-industrial complex.Tim Zadorozhnyy (The Kyiv Independent)
Lemmy has a problem
With a 90% male demographic, Lemmy will face problems related to a homogeneous user population and all the issues that come with it. Right now, it's shaping up to be misogynistic, but it could also head into other bad places. Lemmy needs to attract a more diverse population of users or will end up as another echo chamber for the like minded.
similarweb.com/website/lemmy.m…
don't like this
Wouldn’t a tankie instance have more women?
Your response seems almost defensive, which is weird. Lemmy is definitely overwhelming male. That’s not an inherently bad thing, so I don’t get the defensive tone here or taking OP to task about data methods.
Disagree all you want, but this website is incredibly male dominated. I don’t think OP needs to do a peer-reviewed, double-blind study to say so.
Why are you so aggressively defending a false data collection method?
And what do you mean by "this website"?
The challenge of deleting old online accounts | Loudwhisper
In the last days I spent a disproportionate amount deleting old accounts I found in my password manager, and mostly because so many companies - despite the GDPR - have rudimentary, manually when not completely nonexistent processes to delete your data.
In this post I describe my process going through about 100 old accounts and trying to delete them all, including a top 10 for the weirdest, funniest or most interesting cases I encountered while doing so.
Nice article.
Enjoyed reading it.
A few months ago, I alao went on a small spree of deleting from my ~500 accounts.
Some companies/services were offline, some redirected, some had no or very cumbersome ways to delete my data.
Sometimes I juat wanted to edit my email.
Welp. No can do bro. Your E-Mail is cemented in place and only the heat-death of the universe can remove it.
The ones where you can't edit the email I noticed often used the email as username, and probably god knows how bad is the code on the backend.
Tyler, il figlio maggiore problematico della discussa deputata Lauren Boebert, è accusato di abusi su minori
Il figlio maggiore problematico della deputata Lauren Boebert (cristiana rinata e già sostenitrice della teoria del complotto QAnon) è stato accusato di abuso su minori in seguito a un incidente che ha coinvolto il nipote.
Tyler, il figlio ventenne della deputata repubblicana e dell'ex marito Jayson, è stato accusato di reato minore in Colorado l'11 luglio, ha riferito sabato Denver Westword , citando i registri del dipartimento di polizia di Windsor.
Il deputato Boebert ha minimizzato l'accusa, affermando che era il risultato di "una mancanza di comunicazione sul controllo del mio giovane nipote che di recente lo ha portato ad andarsene di casa".
Rep. Lauren Boebert's troubled eldest son Tyler charged with child abuse
Rep. Boebert downplayed the charge, saying it was the result of “a miscommunication on monitoring my young grandson that recently led to him getting out of our house.”Anthony Blair (New York Post)
How we Rooted Copilot
How we Rooted Copilot - Eye Research
Read how we explored the Python sandbox in Copilot and got root on the underlying containerVaisha Bernard (Eye Security)
"Lasciate che siano i bambini a farlo": il conduttore di Fox News chiede di sostituire gli immigrati con il lavoro minorile
Dopo una visita a una piantagione di mirtilli nel fine settimana, Hurt ha discusso con i conduttori di Fox News Rachel Campos-Duffy e Charlie Kirk sull'opportunità o meno che il governo sovvenzioni le piccole aziende agricole.
"E il dibattito, ragazzi, è su cosa dovrebbe sovvenzionare il governo... voglio dire, guardate, produrre mirtilli richiede molto lavoro, per esempio", ha detto Campos-Duffy. "Quindi cosa dovrebbe sovvenzionare il governo?"
'Allow children to do it': Fox News host calls to replace immigrants with child labor
Fox News host Charlie Hurt argued that President Donald Trump's administration should bring back child labor to replace undocumented immigrants.David Edwards (Raw Story)
Google Gemini deletes use's code
like this
etherphon
in reply to Pro • • •shalafi
in reply to etherphon • • •El_guapazo
in reply to Pro • • •Modern_medicine_isnt
in reply to Pro • • •It's easy, just ask the AI "are you sure"? Until it stops changing it's answer.
But seriously, LLMs are just advanced autocomplete.
Lfrith
in reply to Modern_medicine_isnt • • •GissaMittJobb
in reply to Lfrith • • •jj4211
in reply to GissaMittJobb • • •Also, generally the best interfaces for LLM will combine non-LLM facilities transparently. The LLM might be able to translate the prose to the format the math engine desires and then an intermediate layer recognizes a tag to submit an excerpt to a math engine and substitute the chunk with output from the math engine.
Even for servicing a request to generate an image, the text generation model runs independent of the image generation, and the intermediate layer combines them. Which can cause fun disconnects like the guy asking for a full glass of wine. The text generation half is completely oblivious to the image generation half. So it responds playing the role of a graphic artist dutifully doing the work without ever 'seeing' the image, but it assumes the image is good because that's consistent with training output, but then the user corrects it and it goes about admitting that the picture (that it never 'looked' at) was wrong and retrying the image generator with the additional context, to produce a similarly botched picture.
saimen
in reply to Lfrith • • •jj4211
in reply to Lfrith • • •Modern_medicine_isnt
in reply to jj4211 • • •cley_faye
in reply to Modern_medicine_isnt • • •jj4211
in reply to Modern_medicine_isnt • • •Passerby6497
in reply to jj4211 • • •jj4211
in reply to Passerby6497 • • •It gave me flashbacks when the Replit guy complained that the LLM deleted his data despite being told in all caps not to multiple times.
People really really don't understand how these things work...
Modern_medicine_isnt
in reply to jj4211 • • •Perspectivist
in reply to Pro • • •like this
onewithoutaname likes this.
shalafi
in reply to Perspectivist • • •Neither are our brains.
“Brains are survival engines, not truth detectors. If self-deception promotes fitness, the brain lies. Stops noticing—irrelevant things. Truth never matters. Only fitness. By now you don’t experience the world as it exists at all. You experience a simulation built from assumptions. Shortcuts. Lies. Whole species is agnosiac by default.”
― Peter Watts, Blindsight (fiction)
Starting to think we're really not much smarter. "But LLMs tell us what we want to hear!" Been on FaceBook lately, or lemmy?
If nothing else, LLMs have woke me to how stupid humans are vs. the machines.
Perspectivist
in reply to shalafi • • •like this
onewithoutaname likes this.
aesthelete
in reply to shalafi • • •like this
onewithoutaname likes this.
dontbelievethis
in reply to aesthelete • • •aesthelete
in reply to dontbelievethis • • •saimen
in reply to shalafi • • •jj4211
in reply to shalafi • • •It's not that they may be deceived, it's that they have no concept of what truth or fiction, mistake or success even are.
Our brains know the concepts and may fall to deceipt without recognizing it, but we at least recognize that the concept exists.
An AI generates content that is a blend of material from the training material consistent with extending the given prompt. It only seems to introduce a concept of lying or mistakes when the human injects that into the human half of the prompt material. It will also do so in a way that the human can just as easily instruct it to correct a genuine mistake as well as the human instruct it to correct something that is already correct (unless the training data includes a lot of reaffirmation of the material in the face of such doubts).
An LLM can consume more input than a human can gather in multiple lifetimes and still bo wonky in generating content, because it needs enough to credibly blend content to extend every conceivable input. It's why so many people used to judging human content get derailed by judging AI content. An AI generates a fantastic answer to an interview question that only solid humans get right, only to falter 'on the job' because the utterly generic interview question looks like millions of samples in the input, but the actual job was niche.
Etterra
in reply to Pro • • •Snot Flickerman
in reply to Pro • • •nymnympseudonym
in reply to Snot Flickerman • • •This Nobel Prize winner and subject matter expert takes the opposite view
youtube.com/watch?v=IkdziSLYzH…
- YouTube
youtube.comObinice
in reply to nymnympseudonym • • •People really do not like seeing opposing viewpoints, eh? There's disagreeing, and then there's downvoting to oblivion without even engaging in a discussion, haha.
Even if they're probably right, in such murky uncertain waters where we're not experts, one should have at least a little open mind, or live and let live.
THB
in reply to Obinice • • •It's like talking with someone who thinks the Earth is flat. There isn't anything to discuss. They're objectively wrong.
Humans like to anthropomorphize everything. It's why you can see a face on a car's front grille. LLMs are ultra advanced pattern matching algorithms. They do not think or reason or have any kind of opinion or sentience, yet they are being utilized as if they do. Let's see how it works out for the world, I guess.
saimen
in reply to THB • • •I think so too, but I am really curious what will happen when we give them "bodies" with sensors so they can explore the world and make individual "experiences". I could imagine they would act much more human after a while and might even develop some kind of sentience.
Of course they would also need some kind of memory and self-actualization processes.
jj4211
in reply to saimen • • •Interaction with the physical world isn't really required for us to evaluate how they deal with 'experiences'. They have in principle access to all sorts of interesting experiences in the online data. Some models have been enabled to fetch internet data and add them to the prompt to help synthesize an answer.
One key thing is they don't bother until direction tells them. They don't have any desire they just have "generate search query from prompt, execute search query and fetch results, consider the combination of the original prompt and the results to be the context for generating more content and return to user".
LLM is not a scheme that credibly implies that more LLM == sapient existance. Such a concept may come, but it will be something different than LLM. LLM just looks crazily like dealing with people.
fodor
in reply to Obinice • • •Snot Flickerman
in reply to nymnympseudonym • • •Interesting talk but the number of times he completely dismisses the entire field of linguistics kind of makes me think he's being disingenuous about his familiarity with it.
For one, I think he is dismissing holotes, the concept of "wholeness." That when you cut something apart to it's individual parts, you lose something about the bigger picture. This deconstruction of language misses the larger picture of the human body as a whole, and how every part of us, from our assemblage of organs down to our DNA, impact how we interact with and understand the world. He may have a great definition of understanding but it still sounds (to me) like it's potentially missing aspects of human/animal biologically based understanding.
For example, I have cancer, and about six months before I was diagnosed, I had begun to get more chronically depressed than usual. I felt hopeless and I didn't know why. Surprisingly, that's actually a symptom of my cancer. What understanding did I have that changed how I felt inside and how I understood the things around me? Suddenly I felt different about words and ideas, but nothing had changed externally, something had change internally. The connections in my neural network had adjusted, the feelings and associations with words and ideas was different, but I hadn't done anything to make that adjustment. No learning or understanding had happened. I had a mutation in my DNA that made that adjustment for me.
Further, I think he's deeply misunderstanding (possibly intentionally?) what linguists like Chomsky are saying when they say humans are born with language. They mean that we are born with a genetic blueprint to understand language. Just like animals are born with a genetic blueprint to do things they were never trained to do. Many animals are born and almost immediately stand up to walk. This is the same principle. There are innate biologically ingrained understandings that help us along the path to understanding. It does not mean we are born understanding language as much as we are born with the building blocks of understanding the physical world in which we exist.
Anyway, interesting talk, but I immediately am skeptical of anyone who wholly dismisses an entire field of thought so casually.
For what it's worth, I didn't downvote you and I'm sorry people are doing so.
nymnympseudonym
in reply to Snot Flickerman • • •I am not a linguist but the deafening silence from Chomsky and his defenders really does demand being called out.
Syntactical models of language have been completely crushed by statistics-at-scale via neural nets. But linguists have not rejected the broken model.
The same thing happened with protein folding -- researchers who spent the last 25 years building complex quantum mechanical/electrostatic models of protein structure suddenly saw AlphaFold completely crush prior methods. The difference is, bioinformatics researchers have already done a complete about-face and are .
- YouTube
www.youtube.comgreygore
in reply to nymnympseudonym • • •I watched this entire video just so that I could have an informed opinion. First off, this feels like two very separate talks:
The first part is a decent breakdown of how artificial neural networks process information and store relational data about that information in a vast matrix of numerical weights that can later be used to perform some task. In the case of computer vision, those weights can be used to recognize objects in a picture or video streams, such as whether something is a hotdog or not.
As a side note, if you look up Hinton’s 2024 Nobel Peace Prize in Physics, you’ll see that he won based on his work on the foundations of these neural networks and specifically, their training. He’s definitely an expert on the nuts and bolts about how neural networks work and how to train them.
He then goes into linguistics and how language can be encoded in these neural networks, which is how large language models (LLMs) work… by breaking down words and phrases into tokens and then using the weights in these neural networks to encode how these words relate to each other. These connections are later used to generate other text output related to the text that is used as input. So far so good.
At that point he points out these foundational building blocks have been used to lead to where we are now, at least in a very general sense. He then has what I consider the pivotal slide of the entire talk, labeled Large Language Models, which you can see at 17:22. In particular he has two questions at the bottom of the slide that are most relevant:
* Are they genuinely intelligent?
* Or are they just a form of glorified auto-complete that uses statistical regularities to pastiche together pieces of text that were created by other people?
The problem is: he never answers these questions. He immediately moves on to his own theory about how language works using an analogy to LEGO bricks, and then completely disregards the work of linguists in understanding language, because what do those idiots know?
At this point he brings up The long term existential threat and I would argue the rest of this talk is now science fiction, because it presupposes that understanding the relationship between words is all that is necessary for AI to become superintelligent and therefore a threat to all of us.
Which goes back to the original problem in my opinion: LLMs are text generation machines. They use neural networks encoded as a matrix of weights that can be used to predict long strings of text based on other text. That’s it. You input some text, and it outputs other text based on that original text.
We know that different parts of the brain have different responsibilities. Some parts are used to generate language, other parts store memories, still other parts are used to make our bodies move or regulate autonomous processes like our heartbeat and blood pressure. Still other bits are used to process images from our eyes and other parts reason about spacial awareness, while others engage in emotional regulation and processing.
Saying that having a model for language means that we’ve built an artificial brain is like saying that because I built a round shape called a wheel means that I invented the modern automobile. It’s a small part of a larger whole, and although neural networks can be used to solve some very difficult problems, they’re only a specific tool that can be used to solve very specific tasks.
Although Geoffrey Hinton is an incredibly smart man who mathematically understands neural networks far better than I ever will, extrapolating that knowledge out to believing that a large language model has any kind of awareness or actual intelligence is absurd. It’s the underpants gnome economic theory, but instead of:
1. Collect underpants
2. ?
3. Profit!
It looks more like:
1. Use neural network training to construct large language models.
2. ?
3. Artificial general intelligence!
If LLMs were true artificial intelligence, then they would be learning at an increasing rate as we give them more capacity, leading to the singularity as their intelligence reaches hockey stick exponential growth. Instead, we’ve been throwing a growing amount resources at these LLMs for increasingly smaller returns. We’ve thrown a few extra tricks into the mix, like “reasoning”, but beyond that, I believe it’s clear that we’re headed towards a local maximum that is far enough away from intelligence that would be truly useful (and represent an actual existential threat), but in actuality only resembles what a human can output well enough to fool human decision makers into trusting them to solve problems that they are incapable of solving.
Silicon Valley: Not Hotdog (Season 4 Episode 4 Clip) | HBO
YouTubenymnympseudonym
in reply to greygore • • •I (as a person who works professionally in the area and tries to keep up with the current academic publications) happen to agree with you. But my credences are somewhat reduced after considering the points Hinton raises.
I think it is worth considering that there are a handful of academically active models of consciousness; some well-respected ones like the CTM are not at all inconsistent with Hinton's statements
Lodespawn
in reply to Pro • • •rc__buggy
in reply to Lodespawn • • •Lodespawn
in reply to rc__buggy • • •FanciestPants
in reply to Lodespawn • • •Lodespawn
in reply to FanciestPants • • •Nah so their definition is the classical "how confident are you that you got the answer right". If you read the article they asked a bunch of people and 4 LLMs a bunch of random questions, then asked the respondent whether they/it had confidence their answer was correct, and then checked the answer. The LLMs initially lined up with people (over confident) but then when they iterated, shared results and asked further questions the LLMs confidence increased while people's tends to decrease to mitigate the over confidence.
But the study still assumes intelligence enough to review past results and adjust accordingly, but disregards the fact that an AI isnt intelligence, it's a word prediction model based on a data set of written text tending to infinity. It's not assessing validity of results, it's predicting what the answer is based on all previous inputs. The whole study is irrelevant.
jj4211
in reply to Lodespawn • • •Well, not irrelevant. Lots of our world is trying to treat the LLM output as human-like output, so if human's are going to treat LLM output the same way they treat human generated content, then we have to characterize, for the people, how their expectations are broken in that context.
So as weird as it may seem to treat a stastical content extrapolation engine in the context of social science, there's a great deal of the reality and investment that wants to treat it as "person equivalent" output and so it must be studied in that context, if for no other reason to demonstrate to people that it should be considered "weird".
fodor
in reply to Pro • • •like this
onewithoutaname likes this.
CosmoNova
in reply to Pro • • •kameecoding
in reply to Pro • • •cley_faye
in reply to Pro • • •Oh you.
jj4211
in reply to Pro • • •They are not only unaware of their own mistakes, they are unaware of their successes. They are generating content that is, per their training corpus, consistent with the input. This gets eerie, and the 'uncanny valley' of the mistakes are all the more striking, but they are just generating content without concept of 'mistake' or' 'success' or the content being a model for something else and not just being a blend of stuff from the training data.
For example:
Me: Generate an image of a frog on a lilypad.
LLM: I'll try to create that — a peaceful frog on a lilypad in a serene pond scene. The image will appear shortly below.
Me (lying): That seems to have produced a frog under a lilypad instead of on top.
LLM: Thanks for pointing that out! I'm generating a corrected version now with the frog clearly sitting on top of the lilypad. It’ll appear below shortly.
It didn't know anything about the picture, it just took the input at it's word. A human would have stopped to say "uhh... what do you mean, the lilypad is on water and frog is on top of that?" Or if the human were really trying to just do the request without clarification, they might have tried to think "maybe he wanted it from the perspective of a fish, and he wanted the frog underwater?". A human wouldn't have gone "you are right, I made a mistake, here I've tried again" and include almost the exact same thing.
But tha training data isn't predominantly people blatantly lying about such obvious things or second guessing things that were done so obviously normally correct.
vithigar
in reply to jj4211 • • •The use of language like "unaware" when people are discussing LLMs drives me crazy. LLMs aren't "aware" of anything. They do not have a capacity for awareness in the first place.
People need to stop taking about them using terms that imply thought or consciousness, because it subtly feeds into the idea that they are capable of such.
LainTrain
in reply to vithigar • • •CeeBee_Eh
in reply to Pro • • •This happened to me the other day with Jippity. It outright lied to me:
"You're absolutely right. Although I don't have access to the earlier parts of the conversation".
So it says that I was right in a particular statement, but didn't actually know what I said. So I said to it, you just lied. It kept saying variations of:
"I didn't lie intentionally"
"I understand why it seems that way"
"I wasn't misleading you"
etc
It flat out lied and tried to gaslight me into thinking I was in the wrong for taking that way.
greygore
in reply to CeeBee_Eh • • •It didn’t lie to you or gaslight you because those are things that a person with agency does. Someone who lies to you makes a decision to deceive you for whatever reason they have. Someone who gaslights you makes a decision to behave like the truth as you know it is wrong in order to discombobulate you and make you question your reality.
The only thing close to a decision that LLMs make is: what text can I generate that statistically looks similar to all the other text that I’ve been given. The only reason they answer questions is because in the training data they’ve been provided, questions are usually followed by answers.
It’s not apologizing you to, it knows from its training data that sometimes accusations are followed by language that we interpret as an apology, and sometimes by language that we interpret as pushing back. It regurgitates these apologies without understanding anything, which is why they seem incredibly insincere - it has no ability to be sincere because it doesn’t have any thoughts.
There is no thinking. There are no decisions. The more we anthropomorphize these statistical text generators, ascribing thoughts and feelings and decision making to them, the less we collectively understand what they are, and the more we fall into the trap of these AI marketers about how close we are to truly thinking machines.
CeeBee_Eh
in reply to greygore • • •That's not true. An "if statement" is literally a decision tree.
This is technically true for something like GPT-1. But it hasn't been true for the models trained in the last few years.
It has a large amount of system prompts that alter default behaviour in certain situations. Such as not giving the answer on how to make a bomb. I'm fairly certain there are catches in place to not be overly apologetic to minimize any reputation harm and to reduce potential "liability" issues.
And in that scenario, yes I'm being gaslite because a human told it to.
Partially agree. There's no "thinking" in sentient or sapient sense. But there is thinking in the academic/literal definition sense.
Absolutely false. The entire neural network is billions upon billions of decision trees.
I promise you I know very well what LLMs and other AI systems are. They aren't alive, they do not have human or sapient level of intelligence, and they don't feel. I've actually worked in the AI field for a decade. I've trained countless models. I'm quite familiar with them.
But "gaslighting" is a perfectly fine description of what I explained. The initial conditions were the same and the end result (me knowing the truth and getting irritated about it) were also the same.
greygore
in reply to CeeBee_Eh • • •If you want to engage in a semantically argument, then sure, an “if statement” is a form of decision. This is a worthless distinction that has nothing to do with my original point and I believe you’re aware of that so I’m not sure what this adds to the actual meat of the argument?
Okay, what was added to models trained in the last few years that makes this untrue? To the best of my knowledge, the only advancements have involved:
* Pre-training, which involves some additional steps to add to or modify the initial training data
* Fine-tuning, which is additional training on top of an existing model for specific applications.
* Reasoning, which to the best of my knowledge involves breaking the token output down into stages to give the final output more depth.
* “More”. More training data, more parameters, more GPUs, more power, etc.
I’m hardly an expert in the field, so I could have missed plenty, so what is it that makes it “understand” that a question needs to be answered that doesn’t ultimately go back to the original training data? If I feed it training data that never involves questions, then how will it “know” to answer that question?
System prompts are literally just additional input that is “upstream” of the actual user input, and I fail to see how that changes what I said about it not understanding what an apology is, or how it can be sincere when the LLM is just spitting out words based on their statistical relation to one another?
An LLM doesn’t even understand the concept of right or wrong, much less why lying is bad or when it needs to apologize. It can “apologize” in the sense that it has many examples of apologies that it can synthesize into output when you request one, but beyond that it’s just outputting text. It doesn’t have any understanding of that text.
Again, all that’s doing is adding additional words that can be used in generating output. It’s still just generating text output based on text input. That’s it. It has to know it’s lying or being deceitful in order to gaslight you. Does the text resemble something that can be used to gaslight you? Sure. And if I copy and pasted that from ChatGPT that’s what I’d be doing, but an LLM doesn’t have any real understanding of what it’s outputting so saying that there’s any intent to do anything other than generate text based on other text is just nonsense.
Care to expand on that? Every definition of thinking that I find involves some kind of consideration or reflection, which I would argue that the LLM is not doing, because it’s literally generating output based on a complex system of weighted parameters.
If you want to take the simplest definition of “well, it’s considering what to output and therefore that’s thought”, then I could argue my smart phone is “thinking” because when I tap on a part of the screen it makes decisions about how to respond. But I don’t think anyone would consider that real “thought”.
And a logic gate “decides” what to output. And my lightbulb “decides” whether or not to light up based on the state of the switch. And my alarm “decides” to go off based on what time I set it for last night.
My entire point was to stop anthropomorphizing LLMs by describing what they do as “thought”, and that they don’t make “decisions” in the same way humans do. If you want to use definitions that are overly broad just to say I’m wrong, fine, that’s your prerogative, but it has nothing to do with the idea I was trying to communicate.
Cool.
Sure, if you wanna ascribe human terminology to what marketing companies are calling “artificial intelligence” and further reinforcing misconceptions about how LLMs work, then yeah, you can do that. If you care about people understanding that these algorithms aren’t actually thinking in the same way that humans do, and therefore believing many falsehoods about their capabilities, like I do, then you’d use different terminology.
It’s clear that you don’t care about that and will continue to anthropomorphize these models, so… I guess I’m done here.
melsaskca
in reply to Pro • • •MonkderVierte
in reply to melsaskca • • •jol
in reply to melsaskca • • •As any modern computer system, LLMs are much better and smarter than us at certain tasks while terrible at others. You could say that having good memory and communication skills is part of what defines an intelligent person. Not everyone has those abilities, but LLMs do.
My point is, there's nothing useful coming out of the arguments over the semantics of the word "intelligence".
RoadTrain
in reply to Pro • • •About halfway through the article they quote a paper from 2023:
The LLM space has been changing very quickly over the past few years. Yes, LLMs today still "hallucinate", but you're not doing anyone a service by reporting in 2025 the state of the field over 2 years before.
BeMoreCareful
in reply to Pro • • •Baggie
in reply to Pro • • •Oh god I just figured it out.
It was never they are good at their tasks, faster, or more money efficient.
They are just confident to stupid people.
Christ, it's exactly the same failing upwards that produced the c suite. They've just automated the process.
Snot Flickerman
in reply to Baggie • • •Oh good, so that means we can just replace the C-suite with LLMs then, right? Right?
An AI won't need a Golden Parachute when they inevitably fuck it all up.
kureta
in reply to Pro • • •turmacar
in reply to kureta • • •I find it so incredibly frustrating that we've gotten to the point where the "marketing guys" are not only in charge, but are believed without question, that what they say is true until proven otherwise.
"AI" becoming the colloquial term for LLMs and them being treated as a flawed intelligence instead of interesting generative constructs is purely in service of people selling them as such. And it's maddening. Because they're worthless for that purpose.
REDACTED
in reply to Pro • • •DragonTypeWyvern
in reply to REDACTED • • •REDACTED
in reply to DragonTypeWyvern • • •