• Log InLog In
  • Register
Liquid`
Team Liquid Liquipedia
EDT 08:01
CEST 14:01
KST 21:01
  • Home
  • Forum
  • Calendar
  • Streams
  • Liquipedia
  • Features
  • Store
  • EPT
  • TL+
  • StarCraft 2
  • Brood War
  • Smash
  • Heroes
  • Counter-Strike
  • Overwatch
  • Liquibet
  • Fantasy StarCraft
  • TLPD
  • StarCraft 2
  • Brood War
  • Blogs
Forum Sidebar
Events/Features
News
Featured News
[ASL21] Ro4 Preview: On Course12Code S Season 1 - RO8 Preview7[ASL21] Ro8 Preview Pt2: Progenitors8Code S Season 1 - RO12 Group A: Rogue, Percival, Solar, Zoun13[ASL21] Ro8 Preview Pt1: Inheritors16
Community News
Weekly Cups (May 4-10): Clem, MaxPax, herO win1Maestros of The Game 2 announcement and schedule !10Weekly Cups (April 27-May 4): Clem takes triple0RSL Revival: Season 5 - Qualifiers and Main Event12Code S Season 1 (2026) - RO12 Results1
StarCraft 2
General
MaNa leaves Team Liquid Weekly Cups (May 4-10): Clem, MaxPax, herO win Code S Season 1 - RO8 Preview Behind the Blue - Team Liquid History Book Weekly Cups (April 27-May 4): Clem takes triple
Tourneys
2026 GSL Season 2 Qualifiers $5,000 WardiTV Spring Championship 2026 Maestros of The Game 2 announcement and schedule ! SC2 INu's Battles#16 <BO.9> Master Swan Open (Global Bronze-Master 2)
Strategy
Custom Maps
[D]RTS in all its shapes and glory <3 [A] Nemrods 1/4 players
External Content
Mutation # 525 Wheel of Misfortune The PondCast: SC2 News & Results Mutation # 524 Death and Taxes Mutation # 523 Firewall
Brood War
General
Flashes ASL S21 Ro8 Review BW General Discussion Pros React To: Leta vs Tulbo (ASL S21, Ro.8) (Spoiler) Interview ASL Ro4 Day 2 Winner Data needed
Tourneys
[ASL21] Semifinals A [ASL21] Semifinals B [Megathread] Daily Proleagues [BSL22] RO16 Group Stage - 02 - 10 May
Strategy
Fighting Spirit mining rates [G] Hydra ZvZ: An Introduction Simple Questions, Simple Answers Muta micro map competition
Other Games
General Games
Nintendo Switch Thread Warcraft III: The Frozen Throne Stormgate/Frost Giant Megathread Starcraft Tabletop Miniature Game PC Games Sales Thread
Dota 2
The Story of Wings Gaming
League of Legends
Heroes of the Storm
Simple Questions, Simple Answers Heroes of the Storm 2.0
Hearthstone
Deck construction bug Heroes of StarCraft mini-set
TL Mafia
Vanilla Mini Mafia Mafia Game Mode Feedback/Ideas TL Mafia Community Thread Five o'clock TL Mafia
Community
General
US Politics Mega-thread Russo-Ukrainian War Thread UK Politics Mega-thread YouTube Thread European Politico-economics QA Mega-thread
Fan Clubs
The IdrA Fan Club
Media & Entertainment
[Manga] One Piece Anime Discussion Thread [Req][Books] Good Fantasy/SciFi books
Sports
2024 - 2026 Football Thread McBoner: A hockey love story Formula 1 Discussion
World Cup 2022
Tech Support
streaming software Strange computer issues (software) [G] How to Block Livestream Ads
TL Community
The Automated Ban List
Blogs
How EEG Data Can Predict Gam…
TrAiDoS
ramps on octagon
StaticNine
Funny Nicknames
LUCKY_NOOB
Customize Sidebar...

Website Feedback

Closed Threads



Active: 1444 users

US Politics Mega-thread - Page 5564

Forum Index > General Forum
Post a Reply
Prev 1 5562 5563 5564 5565 5566 5721 Next
Now that we have a new thread, in order to ensure that this thread continues to meet TL standards and follows the proper guidelines, we will be enforcing the rules in the OP more strictly. Be sure to give them a complete and thorough read before posting!

NOTE: When providing a source, please provide a very brief summary on what it's about and what purpose it adds to the discussion. The supporting statement should clearly explain why the subject is relevant and needs to be discussed. Please follow this rule especially for tweets.

Your supporting statement should always come BEFORE you provide the source.


If you have any questions, comments, concern, or feedback regarding the USPMT, then please use this thread: http://www.teamliquid.net/forum/website-feedback/510156-us-politics-thread
Acrofales
Profile Joined August 2010
Spain18291 Posts
Last Edited: 2026-03-17 00:38:23
March 17 2026 00:37 GMT
#111261
On March 17 2026 09:17 LightSpectra wrote:
Show nested quote +
On March 17 2026 09:00 Acrofales wrote:
On March 16 2026 23:00 LightSpectra wrote:
On March 16 2026 15:06 Liquid`Drone wrote:
Your reading comprehension is off. I'm not negative towards ai as a tool for learning and I had no issues with baal posting the summary as a source.

I do have issues with people posting chatgpt posts as arguments but that is different.


LLMs as a learning tool is extremely dubious at best, catastrophic at worst. Aside from the documented fact that they hallucinate up to 40% of their information, the horrific environmental effects, and the predictable outcome of letting the authoritarian billionaire class gatekeep information (remember when Grok would start talking about "white genocide" when asked about literally anything?), they're also extremely sycophantic, which makes overconfident uneducated people even less open to new points of view (Dunning-Kruger effect).

"Horrific environmental effects" is a bit hyperbolic there. Andrew Ng explained it quite well (link) but even if you reject the premise, you're not replacing the compute with nothing, as you would be for AI-slop TikTok videos, but with some other learning aide. I guarantee humans have a bigger carbon footprint than the compute for a few billion tokens a different human would use for education. Let alone the water used by humans, we're horribly inefficient water users.

Don't get me wrong, everything else you mentioned is fine, but the environmental effects is a ridiculous point.


Data centers for AI are projected to use somewhere between 6.7 to 12% of the United States' electricity by 2028: https://www.belfercenter.org/research-analysis/ai-data-centers-us-electric-grid

I somehow doubt non-LLM learning aids are using that much.

Sure. But that's like saying that a lot of electricity is used in lighting. So schools should not have electric lighting.

The amount of electricity used in lighting isn't compared to the total, that's nonsensical. It's compared to the alternative. Electric lighting in schools uses less energy and (more to the point) is less hazardous than gas or candles.

In the case of education, US education in particular, the alternative seems to be to give children a book they won't read and peace out while they watch TikTok videos on their phone. The former is a waste of paper and the latter a greater waste of energy than using an LLM.
LightSpectra
Profile Blog Joined October 2011
United States2575 Posts
March 17 2026 00:49 GMT
#111262
On March 17 2026 09:37 Acrofales wrote:
Show nested quote +
On March 17 2026 09:17 LightSpectra wrote:
On March 17 2026 09:00 Acrofales wrote:
On March 16 2026 23:00 LightSpectra wrote:
On March 16 2026 15:06 Liquid`Drone wrote:
Your reading comprehension is off. I'm not negative towards ai as a tool for learning and I had no issues with baal posting the summary as a source.

I do have issues with people posting chatgpt posts as arguments but that is different.


LLMs as a learning tool is extremely dubious at best, catastrophic at worst. Aside from the documented fact that they hallucinate up to 40% of their information, the horrific environmental effects, and the predictable outcome of letting the authoritarian billionaire class gatekeep information (remember when Grok would start talking about "white genocide" when asked about literally anything?), they're also extremely sycophantic, which makes overconfident uneducated people even less open to new points of view (Dunning-Kruger effect).

"Horrific environmental effects" is a bit hyperbolic there. Andrew Ng explained it quite well (link) but even if you reject the premise, you're not replacing the compute with nothing, as you would be for AI-slop TikTok videos, but with some other learning aide. I guarantee humans have a bigger carbon footprint than the compute for a few billion tokens a different human would use for education. Let alone the water used by humans, we're horribly inefficient water users.

Don't get me wrong, everything else you mentioned is fine, but the environmental effects is a ridiculous point.


Data centers for AI are projected to use somewhere between 6.7 to 12% of the United States' electricity by 2028: https://www.belfercenter.org/research-analysis/ai-data-centers-us-electric-grid

I somehow doubt non-LLM learning aids are using that much.

Sure. But that's like saying that a lot of electricity is used in lighting. So schools should not have electric lighting.


Lighting actually provides a useful service to society.

In the case of education, US education in particular, the alternative seems to be to give children a book they won't read and peace out while they watch TikTok videos on their phone. The former is a waste of paper and the latter a greater waste of energy than using an LLM.


This has got to be one the most false dichotomies ever devised.

"How about instead of LLMs or TikTok we just give them a gun they'll use for the inevitable school shooting they'll do? Saves time at least, and at least this way we get to pick the gun."
2006 Shinhan Bank OSL Season 3 was the greatest tournament of all time
GreenHorizons
Profile Blog Joined April 2011
United States23948 Posts
March 17 2026 01:44 GMT
#111263
On March 17 2026 09:49 LightSpectra wrote:
Show nested quote +
On March 17 2026 09:37 Acrofales wrote:
On March 17 2026 09:17 LightSpectra wrote:
On March 17 2026 09:00 Acrofales wrote:
On March 16 2026 23:00 LightSpectra wrote:
On March 16 2026 15:06 Liquid`Drone wrote:
Your reading comprehension is off. I'm not negative towards ai as a tool for learning and I had no issues with baal posting the summary as a source.

I do have issues with people posting chatgpt posts as arguments but that is different.


LLMs as a learning tool is extremely dubious at best, catastrophic at worst. Aside from the documented fact that they hallucinate up to 40% of their information, the horrific environmental effects, and the predictable outcome of letting the authoritarian billionaire class gatekeep information (remember when Grok would start talking about "white genocide" when asked about literally anything?), they're also extremely sycophantic, which makes overconfident uneducated people even less open to new points of view (Dunning-Kruger effect).

"Horrific environmental effects" is a bit hyperbolic there. Andrew Ng explained it quite well (link) but even if you reject the premise, you're not replacing the compute with nothing, as you would be for AI-slop TikTok videos, but with some other learning aide. I guarantee humans have a bigger carbon footprint than the compute for a few billion tokens a different human would use for education. Let alone the water used by humans, we're horribly inefficient water users.

Don't get me wrong, everything else you mentioned is fine, but the environmental effects is a ridiculous point.


Data centers for AI are projected to use somewhere between 6.7 to 12% of the United States' electricity by 2028: https://www.belfercenter.org/research-analysis/ai-data-centers-us-electric-grid

I somehow doubt non-LLM learning aids are using that much.

Sure. But that's like saying that a lot of electricity is used in lighting. So schools should not have electric lighting.


+ Show Spoiler +
Lighting actually provides a useful service to society.

In the case of education, US education in particular, the alternative seems to be to give children a book they won't read and peace out while they watch TikTok videos on their phone. The former is a waste of paper and the latter a greater waste of energy than using an LLM.

This has got to be one the most false dichotomies ever devised.


"How about instead of LLMs or TikTok we just give them a gun they'll use for the inevitable school shooting they'll do? Saves time at least, and at least this way we get to pick the gun."

That sounds like the Republican counter proposal?
"People like to look at history and think 'If that was me back then, I would have...' We're living through history, and the truth is, whatever you are doing now is probably what you would have done then" "Scratch a Liberal..."
baal
Profile Joined March 2003
10708 Posts
March 17 2026 02:21 GMT
#111264
On March 15 2026 17:59 EnDeR_ wrote:
Show nested quote +
In a lot of parts of the world people can be killed by what they say, also in many parts including the 1st world people can go to prison for the wrong opinions etc.

Manning and Assange went through hell and countless others would without anonymity.


Source for bolded?


[image loading]
Im back, in pog form!
Falling
Profile Blog Joined June 2009
Canada11512 Posts
Last Edited: 2026-03-17 02:27:16
March 17 2026 02:26 GMT
#111265
Oops. Where'd the money go?

Remember all those lawsuits Trump fired out and organizations were complying in advance by settling?
And the settlement money was being funneled into a fund for a future Trump library?

The library fund has been dissolved after it failed to submit an annual report. Another foundation has replaced it with some amount of money but not all of it and no real way to track the sources as there is no mandatory disclosure mechanism.

https://www.ibtimes.co.uk/mystery-missing-millions-trump-library-fund-1785614
ModeratorDavid Duke, Richard Spencer, Nick Fuentes, Daily Stormer... "Some very fine people on both sides"
baal
Profile Joined March 2003
10708 Posts
March 17 2026 02:29 GMT
#111266
On March 15 2026 23:38 LightSpectra wrote:
Show nested quote +
On March 15 2026 20:43 DarkPlasmaBall wrote:
On March 15 2026 19:10 CuddlyCuteKitten wrote:
On March 15 2026 18:18 DarkPlasmaBall wrote:
On March 15 2026 16:44 CuddlyCuteKitten wrote:
On March 15 2026 11:47 DarkPlasmaBall wrote:
On March 15 2026 10:03 LightSpectra wrote:
There shouldn't be anonymity on social media networks, because that's how bad actors from foreign countries and bots sway public opinion. There should absolutely be anonymity for whistleblowers/journalists, purchasing birth control, reading censored literature, etc. But you don't need to do any of those things over social media.

What exactly would you require or not allow, if you're removing anonymity? For example, would you require social media accounts to show first and last names? An accurate, up-to-date profile picture of the person? How would these be verified by the social media platform? And how loose would privacy settings need to be, if a stranger wants to learn more about you and your account? Would stalking or unjustified violence be a concern?

I totally agree with you that bots and bad actors need to be addressed; I'm just not sure how to do that while still protecting average users. (Ideally, tech literacy / social media literacy / research literacy would be taught to new generations in middle/high school, so that young adults are less susceptible to bots, bullshit, and bad actors. That'll take a while though, and something should change during the interim.)


It's really easy. Just require a login with a digital ID or a google/Facebook account created with a digital ID.

We already use it for basically everything, several times per day. Takes social media a few hours to implement.

Doesn't even have to show your full name. Just show country and maybe your first name.
Removes all bots and obvious trolls and obviously illegal shit like childporn.

Ah okay, so you're not talking about removing anonymity between social media posters, but rather simply tying the account to your person behind the scenes.


No real point on complete anonymity, right?
If I want to leave a bad review of my local restaurant I don't want the owner to harass me.
If I want to extort him and other business owners by giving bad reviews the police will find me right away.

The main problem is bots and paid actors pretending to be from other countries, scammers etc. If you can't pretend to be someone completely different, and you have to be a real person that can be tracked down if you commit a crime you solve 99% of the problems and you can still be semi anonymous.

Yeah I think that makes a lot of sense. I'm curious if LightSpectra feels similarly.


I do, yes.

Show nested quote +
On March 15 2026 16:53 baal wrote:
On March 14 2026 22:54 LightSpectra wrote:
Still zero sources so I'll just go ahead and stop bothering.



The fact that you need sources for such a thing speaks volumes.

[image loading]

[image loading]


https://en.wikipedia.org/wiki/Estimated_number_of_civilian_guns_per_capita_by_country



what the hell does that Alabama article have to do with the subject?


You should review the conversation we've had up to this point. I didn't say Cuba didn't have low gun ownership, I'm challenging the notion that IF they had guns they'd definitely overthrow the government because they both want to do that, and no government is capable of defeating a guerilla insurgency of their own citizens.


Unless I"m confusing you with somebody else, you said that it was legal and Cubans had guns and also Venezuens hard armed guerillas as an argument that having armed citizens doesn't help protect against dictatorship.

I never said that its a guaranteed that a citizen guerilla will suceed in overthrowing a government that depends on A LOT of factors, the % of support of the regime, the size of the loyalist army, if there is foreign intervention and many other factors, however is a very strong starting point and it makes society much more resilient against tyranny.

Surely there is a significant tradeoff, there always is, more guns means more guns deaths but I believe the tradeoff is definitely worth it as the random violence that can happen between armed citizens like we see in the US pales in comparison with what a government one day can do.


Im back, in pog form!
Billyboy
Profile Joined September 2024
1719 Posts
March 17 2026 02:38 GMT
#111267
On March 17 2026 11:29 baal wrote:
Show nested quote +
On March 15 2026 23:38 LightSpectra wrote:
On March 15 2026 20:43 DarkPlasmaBall wrote:
On March 15 2026 19:10 CuddlyCuteKitten wrote:
On March 15 2026 18:18 DarkPlasmaBall wrote:
On March 15 2026 16:44 CuddlyCuteKitten wrote:
On March 15 2026 11:47 DarkPlasmaBall wrote:
On March 15 2026 10:03 LightSpectra wrote:
There shouldn't be anonymity on social media networks, because that's how bad actors from foreign countries and bots sway public opinion. There should absolutely be anonymity for whistleblowers/journalists, purchasing birth control, reading censored literature, etc. But you don't need to do any of those things over social media.

What exactly would you require or not allow, if you're removing anonymity? For example, would you require social media accounts to show first and last names? An accurate, up-to-date profile picture of the person? How would these be verified by the social media platform? And how loose would privacy settings need to be, if a stranger wants to learn more about you and your account? Would stalking or unjustified violence be a concern?

I totally agree with you that bots and bad actors need to be addressed; I'm just not sure how to do that while still protecting average users. (Ideally, tech literacy / social media literacy / research literacy would be taught to new generations in middle/high school, so that young adults are less susceptible to bots, bullshit, and bad actors. That'll take a while though, and something should change during the interim.)


It's really easy. Just require a login with a digital ID or a google/Facebook account created with a digital ID.

We already use it for basically everything, several times per day. Takes social media a few hours to implement.

Doesn't even have to show your full name. Just show country and maybe your first name.
Removes all bots and obvious trolls and obviously illegal shit like childporn.

Ah okay, so you're not talking about removing anonymity between social media posters, but rather simply tying the account to your person behind the scenes.


No real point on complete anonymity, right?
If I want to leave a bad review of my local restaurant I don't want the owner to harass me.
If I want to extort him and other business owners by giving bad reviews the police will find me right away.

The main problem is bots and paid actors pretending to be from other countries, scammers etc. If you can't pretend to be someone completely different, and you have to be a real person that can be tracked down if you commit a crime you solve 99% of the problems and you can still be semi anonymous.

Yeah I think that makes a lot of sense. I'm curious if LightSpectra feels similarly.


I do, yes.

On March 15 2026 16:53 baal wrote:
On March 14 2026 22:54 LightSpectra wrote:
Still zero sources so I'll just go ahead and stop bothering.



The fact that you need sources for such a thing speaks volumes.

[image loading]

[image loading]


https://en.wikipedia.org/wiki/Estimated_number_of_civilian_guns_per_capita_by_country



what the hell does that Alabama article have to do with the subject?


You should review the conversation we've had up to this point. I didn't say Cuba didn't have low gun ownership, I'm challenging the notion that IF they had guns they'd definitely overthrow the government because they both want to do that, and no government is capable of defeating a guerilla insurgency of their own citizens.


Unless I"m confusing you with somebody else, you said that it was legal and Cubans had guns and also Venezuens hard armed guerillas as an argument that having armed citizens doesn't help protect against dictatorship.

I never said that its a guaranteed that a citizen guerilla will suceed in overthrowing a government that depends on A LOT of factors, the % of support of the regime, the size of the loyalist army, if there is foreign intervention and many other factors, however is a very strong starting point and it makes society much more resilient against tyranny.

Surely there is a significant tradeoff, there always is, more guns means more guns deaths but I believe the tradeoff is definitely worth it as the random violence that can happen between armed citizens like we see in the US pales in comparison with what a government one day can do.



I going to really stand behind this thing that is already really bad in reality, because there is a small chance it would help against the absolute worst case senario, which is more likely to happen with the really bad thing in reality.
baal
Profile Joined March 2003
10708 Posts
March 17 2026 02:38 GMT
#111268
On March 17 2026 03:46 EnDeR_ wrote:
take AI summaries as facts without checking, like baal was doing, this, to me, is a serious problem.


I was asked for sources for a claim that should be absolutely obvious and common knowledge.

If someone asks for sources that H2O is water I'm not going to make a deep dig into chemistry papers, I was going beyond what anyone should and provided an easy answer for an stupid source request.
Im back, in pog form!
baal
Profile Joined March 2003
10708 Posts
March 17 2026 02:50 GMT
#111269
On March 17 2026 07:37 Liquid`Drone wrote:
Show nested quote +
On March 17 2026 06:50 Dan HH wrote:
On March 17 2026 05:25 Liquid`Drone wrote:
Then - AI used for teaching:
The fact is, AI as it is, is both a shortcut that enables students to quickly get an answer they can hand in - robbing them of any learning they should have gained through the effort of doing the work required to produce that answer - and a fantastic tool for learning for the students who use it correctly. While some people have thought that AI could be like a source of 'democratization' of knowledge, giving everybody access to an assistant that never tires, that can explain stuff in a pedagogical manner, what we've actually seen is that the gap between the 'good' and 'bad' students actually increases because of AI. This isn't just because the bad students are becoming worse (because they ask the chatbot of their choice to answer a question for them and then they get a good answer and think they've done a good job) - but also because the good students are becoming better. Smart students who use AI well, learn more, faster, than they used to do - and it helps them both with relevant facts and connecting the dots.

One of the keys here - and this is an area where education so far has failed entirely, and which is the specific area I'm researching right now - is the importance of prompts. Students - with the occasional exception because they have a teacher with a particular passion for learning about this - have mostly been left to their own with this new tool. But for example, I've supplied my students with the following prompt that they've copy pasted and then used:
+ Show Spoiler +

You should act as a Socratic tutor in social studies for a Vg1 (first-year upper secondary) student.

Topic: What is the most important difference between the Nordic and the Anglo-American welfare model?

I have some prior knowledge, but I need to understand and review the topic better before an assessment.

Rules:

Ask one question at a time

Do not give long explanations

Do not give the full answer unless I ask for it

If my answer is wrong or incomplete, help me move forward with questions or small hints

Ask me to explain things in my own words

After a few rounds, you may give a short summary


And while I don't have a huge dataset or anything, the feedback I got from students - both in terms of them writing a short note containing their reflections on their own learning process - and in terms of them showcasing that they had learned the key differences between the Nordic and Anglo-American welfare model - is highly promising. I've also done a fair amount of work with refugees, and the ones who master the language the fastest tend to be students who use a ton of AI - not 'can you translate this for me please', but 'Can you pretend that you are a native Norwegian? I would like for us to talk about x, and then, when you see that I am frequently making a particular mistake, can you explain what I'm doing wrongly in arabic/ukrainian?'

I'm not some type of oh AI is the savior but man, there's tons of potential - both ways, and I see both, every day, in my working life. And whether we like it or not, it's there, no point in pretending it's not.

I appreciate the time you took to give us a little peek into your work specifically with AI used in learning.

From my own experience, people that were good at googling are good at using AI and vice-versa. The amount of people that are good at googling is low. I'm thinking the percentage of people that are aware search operators exist is a single digit.

With your background, I'm sure you're aware there has been a deluge of studies in the past year comparing depth of learning and cognitive load specifically between groups asked to research a topic using traditional search engines vs LLMs. And the results were invariably that the groups using LLMs used their brains less and retained less information.

Now, we can nitpick methodology, throwing a random subject to someone is one thing, whereas intentionally trying to learn something you're interested in is very different, sure. But we already have the internet itself as a precedent of how people went and used democratized information.

We know the difference between a tool that helps you do the thing and a tool that does the thing for you. Using maps vs GPS is well researched example, GPS makes you not need to look for and memorize landmarks and create a mental map, which hinders your ability to learn to navigate unguided. Not a big deal, unless civilization falls in your lifetime you won't need that ability. Many tasks require skills that you won't need again or that you'll always be assisted with.

But being able to research, interpret and understand and check the veracity of a text are essential skills. And LLMs will hinder them for the average user is my current impression and expectation.

As I said in a previous post, AI isn't inherently bad or sycophantic, there are business decisions making them worse than they need to be. Almost all the ads for AI I've ever seen give banale use cases with some "woah, this thing will think for you" subtext. What's sold to investors is that this thing will do all the science and most of the labor for us and solve climate change and cure cancer, we don't need research grants and education to become more productive, we need hardware farms. What's given to users is compliments and validation and answers that sound confident and authoritative.

To your point that it makes good students better and widening the gap, as we see in the world around us today, it's the median student that decides elections and the future of our species. If the effect of LLMs on the median student is to make them exert their brain less it's a problem. I agree that the solution to that problem isn't to attempt to put the genie back in the box, we're all aware that's not an option and I personally wouldn't take it even if it were an option.


I don't disagree with any of this and I wholeheartedly agree with most of it. The aspect I think I am most worried about with AI is the extension of the GPS example you bring up - we know that our brains actually benefit from doing many of those small, menial tasks we've delegated to our smartphones, and I worry that AI ending up doing more and more of our more creative and rewarding work - not just the small menial tasks - will to a smaller or greater degree rob us of the human experience.

However, I think if you are engaged in the act of discussing a topic and you need some data to prove a point, you're already at a pretty high level of engagement. In that case, I'm not worried about using AI to give a quick data point rather than having to dig around for a reliable and trustworthy source. And as a teacher, the overarching goal of the research team I'm part of is basically to help give teachers around norway the toolbox to help the median student use it wisely.

Also - anecdotally - when ChatGPT was pretty new, I had given a group of students a history test where one of the questions was 'who is considered the father of liberalism'. This question ended up being a fun 'tell' - because the students who had used ChatGPT as their primary tool for preparing for the test all answered John Locke, while the ones who had used our default textbook as their primary tool for preparing answered John Stuart Mill. Either answer is fine (it was a mediocre question, but oh well) - but comparing the rest of the answers, students who went 'John Locke' had generally done a slightly better job overall; without giving me the impression they were harder workers.

This was a class full of good students. Teaching English or civics for vocational students that are going into construction, the overall impression is that AI has been a negative for their learning experience; but even then, I've given assignments that were major hits - that would make students who would normally struggle churning out a single paragraph of their own text over 90 minutes write more than a full page.



Every generation's luddites will have their Encarta, Wikipedia, and now AI moment to make the same mistak

I remember reading an article about it not long ago when book access became popular it was said that it was a lazy way of learning, that its a cheap substitute for lived experience.
Im back, in pog form!
baal
Profile Joined March 2003
10708 Posts
March 17 2026 02:55 GMT
#111270
On March 17 2026 11:38 Billyboy wrote:
I going to really stand behind this thing that is already really bad in reality, because there is a small chance it would help against the absolute worst case senario, which is more likely to happen with the really bad thing in reality.


Perhaps to you in particular an authoritarian regime raising to power is an unlikely scenario, for the majority of the world it is not my friend.


The thing is that the perils of gun ownership scale proportionally with a society's civility, meaning that an "uncivil" country like lets say Mauritania gun ownership creates a lot of violence, but at the same time an authoritarian regime raising to power is very likely, compare that to Switzerland, which has very high gun ownership with little negatives effects but also a Tyranny is very unlikely.

So the risk/reward seems pretty linear across the board.
Im back, in pog form!
EnDeR_
Profile Blog Joined May 2004
Spain2879 Posts
March 17 2026 06:13 GMT
#111271
On March 17 2026 11:21 baal wrote:
Show nested quote +
On March 15 2026 17:59 EnDeR_ wrote:
In a lot of parts of the world people can be killed by what they say, also in many parts including the 1st world people can go to prison for the wrong opinions etc.

Manning and Assange went through hell and countless others would without anonymity.


Source for bolded?


[image loading]


I don't know who that is.
estás más desubicao q un croissant en un plato de nécoras
EnDeR_
Profile Blog Joined May 2004
Spain2879 Posts
March 17 2026 06:25 GMT
#111272
On March 17 2026 11:38 baal wrote:
Show nested quote +
On March 17 2026 03:46 EnDeR_ wrote:
take AI summaries as facts without checking, like baal was doing, this, to me, is a serious problem.


I was asked for sources for a claim that should be absolutely obvious and common knowledge.

If someone asks for sources that H2O is water I'm not going to make a deep dig into chemistry papers, I was going beyond what anyone should and provided an easy answer for an stupid source request.


Providing sources isn't a superfluous request. It's how we figure out what facts you're using when building an argument and making decisions.

Some poster here a while back claimed that they had proof that there had been election interference, and it turned out to be a news report on some election committee that had an emergency meeting near the election.
estás más desubicao q un croissant en un plato de nécoras
Liquid`Drone
Profile Joined September 2002
Norway28797 Posts
March 17 2026 08:31 GMT
#111273
On March 17 2026 11:50 baal wrote:
Show nested quote +
On March 17 2026 07:37 Liquid`Drone wrote:
On March 17 2026 06:50 Dan HH wrote:
On March 17 2026 05:25 Liquid`Drone wrote:
Then - AI used for teaching:
The fact is, AI as it is, is both a shortcut that enables students to quickly get an answer they can hand in - robbing them of any learning they should have gained through the effort of doing the work required to produce that answer - and a fantastic tool for learning for the students who use it correctly. While some people have thought that AI could be like a source of 'democratization' of knowledge, giving everybody access to an assistant that never tires, that can explain stuff in a pedagogical manner, what we've actually seen is that the gap between the 'good' and 'bad' students actually increases because of AI. This isn't just because the bad students are becoming worse (because they ask the chatbot of their choice to answer a question for them and then they get a good answer and think they've done a good job) - but also because the good students are becoming better. Smart students who use AI well, learn more, faster, than they used to do - and it helps them both with relevant facts and connecting the dots.

One of the keys here - and this is an area where education so far has failed entirely, and which is the specific area I'm researching right now - is the importance of prompts. Students - with the occasional exception because they have a teacher with a particular passion for learning about this - have mostly been left to their own with this new tool. But for example, I've supplied my students with the following prompt that they've copy pasted and then used:
+ Show Spoiler +

You should act as a Socratic tutor in social studies for a Vg1 (first-year upper secondary) student.

Topic: What is the most important difference between the Nordic and the Anglo-American welfare model?

I have some prior knowledge, but I need to understand and review the topic better before an assessment.

Rules:

Ask one question at a time

Do not give long explanations

Do not give the full answer unless I ask for it

If my answer is wrong or incomplete, help me move forward with questions or small hints

Ask me to explain things in my own words

After a few rounds, you may give a short summary


And while I don't have a huge dataset or anything, the feedback I got from students - both in terms of them writing a short note containing their reflections on their own learning process - and in terms of them showcasing that they had learned the key differences between the Nordic and Anglo-American welfare model - is highly promising. I've also done a fair amount of work with refugees, and the ones who master the language the fastest tend to be students who use a ton of AI - not 'can you translate this for me please', but 'Can you pretend that you are a native Norwegian? I would like for us to talk about x, and then, when you see that I am frequently making a particular mistake, can you explain what I'm doing wrongly in arabic/ukrainian?'

I'm not some type of oh AI is the savior but man, there's tons of potential - both ways, and I see both, every day, in my working life. And whether we like it or not, it's there, no point in pretending it's not.

I appreciate the time you took to give us a little peek into your work specifically with AI used in learning.

From my own experience, people that were good at googling are good at using AI and vice-versa. The amount of people that are good at googling is low. I'm thinking the percentage of people that are aware search operators exist is a single digit.

With your background, I'm sure you're aware there has been a deluge of studies in the past year comparing depth of learning and cognitive load specifically between groups asked to research a topic using traditional search engines vs LLMs. And the results were invariably that the groups using LLMs used their brains less and retained less information.

Now, we can nitpick methodology, throwing a random subject to someone is one thing, whereas intentionally trying to learn something you're interested in is very different, sure. But we already have the internet itself as a precedent of how people went and used democratized information.

We know the difference between a tool that helps you do the thing and a tool that does the thing for you. Using maps vs GPS is well researched example, GPS makes you not need to look for and memorize landmarks and create a mental map, which hinders your ability to learn to navigate unguided. Not a big deal, unless civilization falls in your lifetime you won't need that ability. Many tasks require skills that you won't need again or that you'll always be assisted with.

But being able to research, interpret and understand and check the veracity of a text are essential skills. And LLMs will hinder them for the average user is my current impression and expectation.

As I said in a previous post, AI isn't inherently bad or sycophantic, there are business decisions making them worse than they need to be. Almost all the ads for AI I've ever seen give banale use cases with some "woah, this thing will think for you" subtext. What's sold to investors is that this thing will do all the science and most of the labor for us and solve climate change and cure cancer, we don't need research grants and education to become more productive, we need hardware farms. What's given to users is compliments and validation and answers that sound confident and authoritative.

To your point that it makes good students better and widening the gap, as we see in the world around us today, it's the median student that decides elections and the future of our species. If the effect of LLMs on the median student is to make them exert their brain less it's a problem. I agree that the solution to that problem isn't to attempt to put the genie back in the box, we're all aware that's not an option and I personally wouldn't take it even if it were an option.


I don't disagree with any of this and I wholeheartedly agree with most of it. The aspect I think I am most worried about with AI is the extension of the GPS example you bring up - we know that our brains actually benefit from doing many of those small, menial tasks we've delegated to our smartphones, and I worry that AI ending up doing more and more of our more creative and rewarding work - not just the small menial tasks - will to a smaller or greater degree rob us of the human experience.

However, I think if you are engaged in the act of discussing a topic and you need some data to prove a point, you're already at a pretty high level of engagement. In that case, I'm not worried about using AI to give a quick data point rather than having to dig around for a reliable and trustworthy source. And as a teacher, the overarching goal of the research team I'm part of is basically to help give teachers around norway the toolbox to help the median student use it wisely.

Also - anecdotally - when ChatGPT was pretty new, I had given a group of students a history test where one of the questions was 'who is considered the father of liberalism'. This question ended up being a fun 'tell' - because the students who had used ChatGPT as their primary tool for preparing for the test all answered John Locke, while the ones who had used our default textbook as their primary tool for preparing answered John Stuart Mill. Either answer is fine (it was a mediocre question, but oh well) - but comparing the rest of the answers, students who went 'John Locke' had generally done a slightly better job overall; without giving me the impression they were harder workers.

This was a class full of good students. Teaching English or civics for vocational students that are going into construction, the overall impression is that AI has been a negative for their learning experience; but even then, I've given assignments that were major hits - that would make students who would normally struggle churning out a single paragraph of their own text over 90 minutes write more than a full page.



Every generation's luddites will have their Encarta, Wikipedia, and now AI moment to make the same mistak

I remember reading an article about it not long ago when book access became popular it was said that it was a lazy way of learning, that its a cheap substitute for lived experience.


This one isn't necessarily unfounded though. We do know that smartphones have negatively affected youth in many ways - especially in terms of concentrating and maintaining focus for a prolonged period of time - and this is confirmed by neurologists. There's also research showing that brain activity is reduced for students who use chatGPT to write papers for them - but increases for ones who wrote their papers independently first and then use LLMs later. But at least in Norway, to what degree teachers know about this stuff is highly dependent on the individual teacher - and the policy of the school they work at, and I suspect it is the same way other places.
Moderator
Liquid`Drone
Profile Joined September 2002
Norway28797 Posts
March 17 2026 08:41 GMT
#111274
On March 17 2026 06:34 EnDeR_ wrote:
Show nested quote +
On March 17 2026 05:36 Liquid`Drone wrote:
On March 17 2026 03:46 EnDeR_ wrote:
On March 17 2026 02:06 Liquid`Drone wrote:
On March 16 2026 23:00 LightSpectra wrote:
On March 16 2026 15:06 Liquid`Drone wrote:
Your reading comprehension is off. I'm not negative towards ai as a tool for learning and I had no issues with baal posting the summary as a source.

I do have issues with people posting chatgpt posts as arguments but that is different.


LLMs as a learning tool is extremely dubious at best, catastrophic at worst. Aside from the documented fact that they hallucinate up to 40% of their information, the horrific environmental effects, and the predictable outcome of letting the authoritarian billionaire class gatekeep information (remember when Grok would start talking about "white genocide" when asked about literally anything?), they're also extremely sycophantic, which makes overconfident uneducated people even less open to new points of view (Dunning-Kruger effect).


Grok is an outlier and should not be trusted for anything.

I can also be on board with being negative towards using ChatGPT because OpenAI - unlike Anthropic, didn't refuse to cooperate with the pentagon regarding mass surveillance or fully autonomous weapons. I'm not gonna argue against the environmental effects, but this idea that AI is entirely bad and without positive sides is nonsense.

If you talk to chatgpt or copilot or gemini about subjects that are uncontroversial or well established, it's good. You'll get solid answers that largely correspond with the truth (or well, the 'most accepted/established information'). When you say they 'hallucinate up to 40% of their information', what does that even mean? You think it's wrong 40% of the time? Or that 'on certain weird, niche subjects where it doesn't have much knowledge, it will still pretend to know what it's talking about and then it can, in those specific situations, confidently make up 40% of what it tells you'? That's an issue - for sure - but if you want to educate yourself on photosyntesis or grammar rules or the consequences of the industrial revolution, LLMs are fantastic.


There are many upsides of using genAI tools. I use it regularly. I should also say that in a scientific context, I find the AI to be more inaccurate than accurate, no matter the model I'm using. Granted, I'm not asking how photosynthesis works.

Anecdotally, AI summaries is how many students learn; why bother with the course material when you can just study a summary of it.

My PhD students don't read papers, they read AI summaries of papers. In a scientific context, this is bad because to make the AI summary, it's dumbing down the content and giving results that are inaccurate. My students can't tell that they're getting inaccurate information and this is becoming a serious problem.

My point was that AI summaries, especially within search results, has a huge problem: either the tech company decides which information you see (by biasing sources), or it feeds you whatever was in the search results, so you have no way of telling (unless you go and check the sources!) whether the information is coming from a reputable source or not. When people unquestionably take AI summaries as facts without checking, like baal was doing, this, to me, is a serious problem.


My wife is doing a PhD and for her, AI has been an invaluable asset. In particular for statistical analysis which she herself did not have the skillset to do. Then, she's gotten stuff double and triple-checked by people with those skills, and overwhelmingly, AI has been spot on. She doesn't use it the way your students do - and I agree it's important that students develop the grit and tenacity to actually read academic papers - but if I want to read a PhD paper about a scientific field where I'm not literate, I will get a more accurate understanding from reading an AI summary than I will from reading the actual paper. If it's a field where I know my shit, nope, still gonna trust myself.

For the AI summary part - it's basically like googling something, visiting the first link you get, and using data from that link as your evidence. Can it be wrong? Sure. But it's a) significantly less time consuming and b) not wrong by default especially if you're talking about a non-controversial topic. Like baal himself said, using an AI summary to get an answer to "Does the cuban people hate their government" would be awful - but for 'gun ownership rates for different countries', it's fine.


Agree that AI can help you do stuff that you don't have the background for. Especially if the solution involves writing some code (like doing statistical analysis). This is fine if it's a well-defined problem with known solutions. It quickly craps out if you want to apply it to something new but I digress.

To your latter point, this isn't as simple as you are making it out to be. A question like "how many immigrants did ICE deport in 2025" should not be controversial, it should just be a number collected from reports, similar to how numbers for gun ownership for different countries are collected. And yet, would you trust any number that it gives you? Would you trust it if it summarised the white house website? I am not saying that it's wrong by default, but that you can't assume the answer is right without checking it. And people don't check. I don't think this is a good direction of travel.



But the AI summary does give sources. Incidentally - for whatever reason - when I just googled 'how many immigrants did ICE deport in 2025' - there wasn't even an AI summary created for me, and I just got a bunch of different sources. However, asking ChatGPT that very same question, I get a pretty nuanced answer - which links to four different sources:

+ Show Spoiler +
The exact number is uncertain, because official U.S. data for 2025 has been incomplete and inconsistent. But based on the best available estimates:

≈ 540,000 people were deported in 2025 (widely cited estimate) source

Other analyses/projected totals suggest roughly 500,000–600,000 deportations for the year (source

Some government-related figures and reporting also point to totals in that same range (around 540,000) (source

Important context


This total usually includes deportations carried out by ICE + border authorities (CBP), not just ICE alone.

Estimates vary because:

The government released limited or inconsistent data in 2025 source

Different sources count deportations slightly differently (e.g., interior removals vs. border removals).

Bottom line

👉 A reasonable, evidence-based answer is:
About 500,000–600,000 immigrants were deported from the U.S. in 2025, with ~540,000 being the most commonly cited figure.


In my opinion this is a very solid answer to the question I asked. It's not overconfident, and it uses a reasonable set of sources as its foundation for its answer. And this corresponds very well with my experience using chatgpt or copilot as a google-substitute - or even looking at the AI summary. When I wrote my previous post where I linked baal a source from Education Weekly about how ai use influences brain activity, that was a source I got from the AI summary of my google search.
Moderator
Velr
Profile Blog Joined July 2008
Switzerland10884 Posts
Last Edited: 2026-03-17 09:43:57
March 17 2026 09:38 GMT
#111275
On March 17 2026 11:55 baal wrote:
Show nested quote +
On March 17 2026 11:38 Billyboy wrote:
I going to really stand behind this thing that is already really bad in reality, because there is a small chance it would help against the absolute worst case senario, which is more likely to happen with the really bad thing in reality.


Perhaps to you in particular an authoritarian regime raising to power is an unlikely scenario, for the majority of the world it is not my friend.


The thing is that the perils of gun ownership scale proportionally with a society's civility, meaning that an "uncivil" country like lets say Mauritania gun ownership creates a lot of violence, but at the same time an authoritarian regime raising to power is very likely, compare that to Switzerland, which has very high gun ownership with little negatives effects but also a Tyranny is very unlikely.

So the risk/reward seems pretty linear across the board.



In Switzerland "we" got guns to defend "our goverment" because our goverment is us/our community.
I never heard from anyone owning guns to protect himself from "our goverment" or goverment overreach or whatever fantasy US Citizens like to tell themselves. Atleast outside of some batshit crazy ultra right wing "reichsbürger/free citizen" types.
baal
Profile Joined March 2003
10708 Posts
March 17 2026 10:14 GMT
#111276
On March 17 2026 15:13 EnDeR_ wrote:
Show nested quote +
On March 17 2026 11:21 baal wrote:
On March 15 2026 17:59 EnDeR_ wrote:
In a lot of parts of the world people can be killed by what they say, also in many parts including the 1st world people can go to prison for the wrong opinions etc.

Manning and Assange went through hell and countless others would without anonymity.


Source for bolded?


[image loading]


I don't know who that is.


Count Dankula, arrested and went to court for a nazi joke with his dog, google him.

Im back, in pog form!
baal
Profile Joined March 2003
10708 Posts
March 17 2026 10:18 GMT
#111277
On March 17 2026 17:31 Liquid`Drone wrote:
Show nested quote +
On March 17 2026 11:50 baal wrote:
On March 17 2026 07:37 Liquid`Drone wrote:
On March 17 2026 06:50 Dan HH wrote:
On March 17 2026 05:25 Liquid`Drone wrote:
Then - AI used for teaching:
The fact is, AI as it is, is both a shortcut that enables students to quickly get an answer they can hand in - robbing them of any learning they should have gained through the effort of doing the work required to produce that answer - and a fantastic tool for learning for the students who use it correctly. While some people have thought that AI could be like a source of 'democratization' of knowledge, giving everybody access to an assistant that never tires, that can explain stuff in a pedagogical manner, what we've actually seen is that the gap between the 'good' and 'bad' students actually increases because of AI. This isn't just because the bad students are becoming worse (because they ask the chatbot of their choice to answer a question for them and then they get a good answer and think they've done a good job) - but also because the good students are becoming better. Smart students who use AI well, learn more, faster, than they used to do - and it helps them both with relevant facts and connecting the dots.

One of the keys here - and this is an area where education so far has failed entirely, and which is the specific area I'm researching right now - is the importance of prompts. Students - with the occasional exception because they have a teacher with a particular passion for learning about this - have mostly been left to their own with this new tool. But for example, I've supplied my students with the following prompt that they've copy pasted and then used:
+ Show Spoiler +

You should act as a Socratic tutor in social studies for a Vg1 (first-year upper secondary) student.

Topic: What is the most important difference between the Nordic and the Anglo-American welfare model?

I have some prior knowledge, but I need to understand and review the topic better before an assessment.

Rules:

Ask one question at a time

Do not give long explanations

Do not give the full answer unless I ask for it

If my answer is wrong or incomplete, help me move forward with questions or small hints

Ask me to explain things in my own words

After a few rounds, you may give a short summary


And while I don't have a huge dataset or anything, the feedback I got from students - both in terms of them writing a short note containing their reflections on their own learning process - and in terms of them showcasing that they had learned the key differences between the Nordic and Anglo-American welfare model - is highly promising. I've also done a fair amount of work with refugees, and the ones who master the language the fastest tend to be students who use a ton of AI - not 'can you translate this for me please', but 'Can you pretend that you are a native Norwegian? I would like for us to talk about x, and then, when you see that I am frequently making a particular mistake, can you explain what I'm doing wrongly in arabic/ukrainian?'

I'm not some type of oh AI is the savior but man, there's tons of potential - both ways, and I see both, every day, in my working life. And whether we like it or not, it's there, no point in pretending it's not.

I appreciate the time you took to give us a little peek into your work specifically with AI used in learning.

From my own experience, people that were good at googling are good at using AI and vice-versa. The amount of people that are good at googling is low. I'm thinking the percentage of people that are aware search operators exist is a single digit.

With your background, I'm sure you're aware there has been a deluge of studies in the past year comparing depth of learning and cognitive load specifically between groups asked to research a topic using traditional search engines vs LLMs. And the results were invariably that the groups using LLMs used their brains less and retained less information.

Now, we can nitpick methodology, throwing a random subject to someone is one thing, whereas intentionally trying to learn something you're interested in is very different, sure. But we already have the internet itself as a precedent of how people went and used democratized information.

We know the difference between a tool that helps you do the thing and a tool that does the thing for you. Using maps vs GPS is well researched example, GPS makes you not need to look for and memorize landmarks and create a mental map, which hinders your ability to learn to navigate unguided. Not a big deal, unless civilization falls in your lifetime you won't need that ability. Many tasks require skills that you won't need again or that you'll always be assisted with.

But being able to research, interpret and understand and check the veracity of a text are essential skills. And LLMs will hinder them for the average user is my current impression and expectation.

As I said in a previous post, AI isn't inherently bad or sycophantic, there are business decisions making them worse than they need to be. Almost all the ads for AI I've ever seen give banale use cases with some "woah, this thing will think for you" subtext. What's sold to investors is that this thing will do all the science and most of the labor for us and solve climate change and cure cancer, we don't need research grants and education to become more productive, we need hardware farms. What's given to users is compliments and validation and answers that sound confident and authoritative.

To your point that it makes good students better and widening the gap, as we see in the world around us today, it's the median student that decides elections and the future of our species. If the effect of LLMs on the median student is to make them exert their brain less it's a problem. I agree that the solution to that problem isn't to attempt to put the genie back in the box, we're all aware that's not an option and I personally wouldn't take it even if it were an option.


I don't disagree with any of this and I wholeheartedly agree with most of it. The aspect I think I am most worried about with AI is the extension of the GPS example you bring up - we know that our brains actually benefit from doing many of those small, menial tasks we've delegated to our smartphones, and I worry that AI ending up doing more and more of our more creative and rewarding work - not just the small menial tasks - will to a smaller or greater degree rob us of the human experience.

However, I think if you are engaged in the act of discussing a topic and you need some data to prove a point, you're already at a pretty high level of engagement. In that case, I'm not worried about using AI to give a quick data point rather than having to dig around for a reliable and trustworthy source. And as a teacher, the overarching goal of the research team I'm part of is basically to help give teachers around norway the toolbox to help the median student use it wisely.

Also - anecdotally - when ChatGPT was pretty new, I had given a group of students a history test where one of the questions was 'who is considered the father of liberalism'. This question ended up being a fun 'tell' - because the students who had used ChatGPT as their primary tool for preparing for the test all answered John Locke, while the ones who had used our default textbook as their primary tool for preparing answered John Stuart Mill. Either answer is fine (it was a mediocre question, but oh well) - but comparing the rest of the answers, students who went 'John Locke' had generally done a slightly better job overall; without giving me the impression they were harder workers.

This was a class full of good students. Teaching English or civics for vocational students that are going into construction, the overall impression is that AI has been a negative for their learning experience; but even then, I've given assignments that were major hits - that would make students who would normally struggle churning out a single paragraph of their own text over 90 minutes write more than a full page.



Every generation's luddites will have their Encarta, Wikipedia, and now AI moment to make the same mistak

I remember reading an article about it not long ago when book access became popular it was said that it was a lazy way of learning, that its a cheap substitute for lived experience.


This one isn't necessarily unfounded though. We do know that smartphones have negatively affected youth in many ways - especially in terms of concentrating and maintaining focus for a prolonged period of time - and this is confirmed by neurologists. There's also research showing that brain activity is reduced for students who use chatGPT to write papers for them - but increases for ones who wrote their papers independently first and then use LLMs later. But at least in Norway, to what degree teachers know about this stuff is highly dependent on the individual teacher - and the policy of the school they work at, and I suspect it is the same way other places.


None were completely unfounded, just nearsighted, some students copy-pasted wikipedia, many teachers wanted to force students to use the library, then later realized the way information worked just changed.
Im back, in pog form!
baal
Profile Joined March 2003
10708 Posts
Last Edited: 2026-03-17 10:23:40
March 17 2026 10:23 GMT
#111278
On March 17 2026 18:38 Velr wrote:
Show nested quote +
On March 17 2026 11:55 baal wrote:
On March 17 2026 11:38 Billyboy wrote:
I going to really stand behind this thing that is already really bad in reality, because there is a small chance it would help against the absolute worst case senario, which is more likely to happen with the really bad thing in reality.


Perhaps to you in particular an authoritarian regime raising to power is an unlikely scenario, for the majority of the world it is not my friend.


The thing is that the perils of gun ownership scale proportionally with a society's civility, meaning that an "uncivil" country like lets say Mauritania gun ownership creates a lot of violence, but at the same time an authoritarian regime raising to power is very likely, compare that to Switzerland, which has very high gun ownership with little negatives effects but also a Tyranny is very unlikely.

So the risk/reward seems pretty linear across the board.



In Switzerland "we" got guns to defend "our goverment" because our goverment is us/our community.
I never heard from anyone owning guns to protect himself from "our goverment" or goverment overreach or whatever fantasy US Citizens like to tell themselves. Atleast outside of some batshit crazy ultra right wing "reichsbürger/free citizen" types.


Yes as I said, Switzerland has a very low chance of falling into tyranny, and also has very little societal issues with high gun ownership so it would be silly to be worried about tyranny, Rwanda on the other hand...
Im back, in pog form!
Uldridge
Profile Blog Joined January 2011
Belgium5158 Posts
Last Edited: 2026-03-17 11:01:37
March 17 2026 10:41 GMT
#111279
Or Russia, or the US, or China. No need to use a small player like Rwamda. Russia and China are important examples because they care so little about human rights they'll just outright kill protestors as a deterrent. And they'll do it over and over again up to the point the masses become so enraged and emboldened they topple the government. Hasn't happened in the US yet as far as I know, but who knows what will happen in the near future.
Taxes are for Terrans
EnDeR_
Profile Blog Joined May 2004
Spain2879 Posts
March 17 2026 11:42 GMT
#111280
On March 17 2026 19:14 baal wrote:
Show nested quote +
On March 17 2026 15:13 EnDeR_ wrote:
On March 17 2026 11:21 baal wrote:
On March 15 2026 17:59 EnDeR_ wrote:
In a lot of parts of the world people can be killed by what they say, also in many parts including the 1st world people can go to prison for the wrong opinions etc.

Manning and Assange went through hell and countless others would without anonymity.


Source for bolded?


[image loading]


I don't know who that is.


Count Dankula, arrested and went to court for a nazi joke with his dog, google him.



On 23 April 2018, Meechan was sentenced to a fine of £800, with no prison sentence.[11]

so not going to prison for his opinions?
estás más desubicao q un croissant en un plato de nécoras
Prev 1 5562 5563 5564 5565 5566 5721 Next
Please log in or register to reply.
Live Events Refresh
Kung Fu Cup
11:00
#7
IntoTheiNu 1171
RotterdaM329
WardiTV229
TKL 188
SteadfastSC59
Liquipedia
Replay Cast
09:00
KungFu Cup 2026 Week 6
CranKy Ducklings132
LiquipediaDiscussion
[ Submit Event ]
Live Streams
Refresh
StarCraft 2
RotterdaM 323
TKL 188
ProTech146
Rex 85
SteadfastSC 59
herO (SOOP) 17
StarCraft: Brood War
Britney 35337
Calm 8737
Sea 4391
Bisu 1560
Jaedong 646
Horang2 578
firebathero 507
Soma 453
actioN 407
Hyuk 319
[ Show more ]
Mini 254
Killer 230
EffOrt 200
Last 166
Pusan 143
Zeus 131
Mind 105
ZerO 101
Rush 91
Larva 78
ggaemo 65
Aegong 49
Soulkey 48
HiyA 44
sSak 44
ToSsGirL 44
Sharp 42
hero 36
Hm[arnc] 30
JulyZerg 26
sorry 24
soO 16
Icarus 14
Bale 13
Movie 11
Noble 9
ajuk12(nOOB) 9
Sexy 7
IntoTheRainbow 7
Terrorterran 5
Dota 2
Gorgc4998
XcaliburYe160
Counter-Strike
olofmeister2674
shoxiejesuss563
x6flipin445
edward70
kRYSTAL_13
Other Games
singsing1309
B2W.Neo498
DeMusliM240
XaKoH 223
Lowko213
Beastyqt175
monkeys_forever116
Mew2King101
amsayoshi34
ZerO(Twitch)8
Happy0
Organizations
Counter-Strike
PGL34151
StarCraft: Brood War
UltimateBattle 131
lovetv 11
StarCraft 2
Blizzard YouTube
StarCraft: Brood War
BSLTrovo
[ Show 16 non-featured ]
StarCraft 2
• CranKy Ducklings SOOP28
• intothetv
• AfreecaTV YouTube
• Kozan
• IndyKCrew
• LaughNgamezSOOP
• Migwel
• sooper7s
StarCraft: Brood War
• iopq 9
• BSLYoutube
• STPLYoutube
• ZZZeroYoutube
Dota 2
• lizZardDota259
League of Legends
• Nemesis4294
• Jankos1622
Other Games
• WagamamaTV226
Upcoming Events
Replay Cast
12h
The PondCast
22h
OSC
22h
Replay Cast
1d 12h
RSL Revival
1d 22h
OSC
2 days
Korean StarCraft League
2 days
RSL Revival
2 days
BSL
3 days
GSL
3 days
Cure vs herO
SHIN vs Maru
[ Show More ]
BSL
4 days
Replay Cast
4 days
Monday Night Weeklies
5 days
Replay Cast
5 days
The PondCast
5 days
GSL
6 days
Liquipedia Results

Completed

Proleague 2026-05-12
WardiTV TLMC #16
Nations Cup 2026

Ongoing

BSL Season 22
ASL Season 21
IPSL Spring 2026
KCM Race Survival 2026 Season 2
Acropolis #4
KK 2v2 League Season 1
BSL 22 Non-Korean Championship
SCTL 2026 Spring
RSL Revival: Season 5
2026 GSL S1
Asian Champions League 2026
IEM Atlanta 2026
PGL Astana 2026
BLAST Rivals Spring 2026
IEM Rio 2026
PGL Bucharest 2026
Stake Ranked Episode 1
BLAST Open Spring 2026
ESL Pro League S23 Finals
ESL Pro League S23 Stage 1&2

Upcoming

Escore Tournament S2: W7
YSL S3
Escore Tournament S2: W8
CSLAN 4
Kung Fu Cup 2026 Grand Finals
HSC XXIX
uThermal 2v2 2026 Main Event
Maestros of the Game 2
WardiTV Spring 2026
2026 GSL S2
BLAST Bounty Summer 2026: Closed Qualifier
Stake Ranked Episode 3
XSE Pro League 2026
IEM Cologne Major 2026
Stake Ranked Episode 2
CS Asia Championships 2026
TLPD

1. ByuN
2. TY
3. Dark
4. Solar
5. Stats
6. Nerchio
7. sOs
8. soO
9. INnoVation
10. Elazer
1. Rain
2. Flash
3. EffOrt
4. Last
5. Bisu
6. Soulkey
7. Mini
8. Sharp
Sidebar Settings...

Advertising | Privacy Policy | Terms Of Use | Contact Us

Original banner artwork: Jim Warren
The contents of this webpage are copyright © 2026 TLnet. All Rights Reserved.