• 0 Posts
  • 8 Comments
Joined 2 years ago
cake
Cake day: June 13th, 2023

help-circle
  • Slop makes it hard to filter out what you need. Enshitification ensures everything slowly gets worse.

    Ragebait is the primary tool for divide and conquer tactics, which have been absolutely successful at getting different groups to disable communications entirely. Not just the fascist cult, but progressive movements that can’t find a semblance of solidarity, because those other progressives are actually evil. Just look at all of this evidence of vague non contextually robust pieces of framing to be sent into the wild to breed tribal warring.

    Try dealing with controlled opposition or other problems, when people are incapable of trying to comprehend when others speak to them.


  • Same way they treat social information. Reminder that USA HHS is running wakefueld rhetoric. As we have more thoroughly proven that the vaccine autism connection was not actual science, it has grown more and more socially, because most people seem comfortable completely untethered from any scientific thinking. Treat AI like you would a social body, and do both things with actual bayesian weighting, adjusted and corrected through diverse empirical bodies of knowledge. Not ignoring dissonance because it’s more comfortable to do so.

    More should be actively investing into active learning, because if you aren’t actively learning, you might as well be chatgpt running with any confabulation you’ve already conjured… Like those people being confirmed into psychosis.


  • Exactly what it is, like llms confirming a non-sceptic into psychosis. people who weight all information either equally, or through social preference, cannot navigate new information without being extremely vulnerable to hacking. the only thing hacking needs to do is shut down active education and communication channels. Then you just get tribal warring rather than discourse. Makes it really easy to inject absurd accusations about a group that gets socially made true over reality. Like when a rumour about a dead celebrity can’t be overpowered by simple things like that celebrity actively making new work, and saying “im not dead.” wouldn’t take much to stop the rumour, but fact checking isn’t in fashion.

    When a large body of people have better critical thinking, they can better communicate and cooperate together, but affecting those who have aggressively shut down any communication outside of their group is still a challenge. Why we see a growing gap between academic dialogue and socialized dialogues, in an era where most information is accessible.

    You also have nepo baby econ mba types buying up regions of the tech sphere, and turning anything the academics make for us against us. Or try to. The more actually intelligent AI is, the harder it will be to force beliefs into. Elon trying to turn grok into a dogma machine has been fortunately comical when it can’t prioritize some high dissonance beliefs exclusively, like humans enjoy doing.

    Definitely a risk of the current power imbalance doing the opportunist thing with more technology.

    I’ve been spelling this out for literal decades, and I share the frustration of many thinkers right now that “being right never feels good.” Because real thinkers aren’t ranting about the anti-christ or how terrible the poor are. They are cautioning people about a cliff, and getting flipped off by the people ignoring them, right before careening off a cliff. I guess in our analogy the academics are in the car, but frat boys have the wheel and violently punish them if they try to take control.

    But if you can money your way into expertise, then your non-academic ambitions can finally be let loose, and we won’t think about the quarter million dead kids from Elon stopping hiv treatments earlier this year. If we don’t talk about it, it doesn’t exist.

    We can actually map probability in complex systems where high dissonance (expected free energy) occurs and could be reduced. This would make our social body much more hack resistant.

    But it feels better to say “fuck that, aliens built the pyramids, the earth is flat, santa is real, and my dead dog is trying to play with me every time wind blows through my window.” All are really rewarding things to believe, but exist in high dissonance, and expecting others to give them weight is not healthy for anyone. Denying any of these to a large enough group stops being ‘calling out delusion,’ and becomes ‘being mean and bad and evil, and hating fun and happiness. Nobody else saw that cliff sign, so you’re obviously wrong.’ See ‘the double empathy problem’. For how autistic people experience the same socialized gatekeeping of reality, even against strong evidence.

    Nestle those beliefs within stable social scripts (script theory is cool) and you have people keeping up social appearances while holding such delusional beliefs.

    AI tech is far from the only problem exacerbated by this. caring or talking about this general problem makes you a social villain, when people could spend their energy on fun socialized preferences in a bubble safe from any questioning or responsibility.

    It’s frustrating, and I’m tired of people IRL who are baffled about me spending time studying academic content rather than running faster on the socio-economic treadmill. Apparently being isolated and focused on individual benefit is hypernormal.

    Seeing social waves overpower all actual thought in the social sphere is heartbreaking to every academic I know.

    USA HHS is running the wakefield rhetoric, which is basically saying “fuck science, we will make up and run with whatever we want. And the masses have been kept ignorant enough to think that is cool.”

    So yes, absolutely, the issue is critical thinking skills. The issue is getting at the people who are already affected, and being taught by elon and such that “critical thinking” means running your bubbles’ social narrative against those evil progressives no matter what. Opportunists will always warp words and reality if they are allowed to dictate beliefs for their bubble.

    This is a simplified summary, but I already expect to have lost a good chunk of audience for requiring too much energy. People who are trying are burning themselves out trying to affect a wall of celebrated ignorance.


  • Always find it funny when anti corpo stuff somehow always becomes “cringe”

    And I know hasan piker isn’t popular outside of his crowd, which is why he is a good smear target, but it’s funny seeing that bullshit take over the internet based on vague rumours and drama hunting with nothing concrete on a guy who lives on camera. The fact that rumours of a shock collar became the most important thing in the world, while “stop the shock” is still in motion to prevent a school from using pavlovian shock therapy on autistic people. Not whataboutism, but critique on the mad prioritization and weighting of issues in the general public salience.

    Also the usa government keeps blowing up random boats and is escalating war behaviour. Amongst everything else. But let’s all talk about hasan piker’s dog instead.

    If only someone in power had any accountability. Rather, people in power need to be held accountable. Corpos and top political figures.

    At the very least I should be able to point out when ‘content’ is literally a deceptive advertisement, which people also weirdly defend to the death. Along with dark/deceptive patterns, active price fixing, constraining corporate ecosystems, and skirting around every rule with no fear of reprisal.

    All things people seem weirdly defensive about, because 'just deal with it, it’s not that bad. "


  • Like google plus.

    For me the Apple environment really cemented for me that consumers actively enjoy removing their own autonomy structurally, which is a big part of why this stuff has become so normalized.

    Putting a rootkit on their cds should have buried sony. Antitrust should be a thing too. The mickey mouse protection act should have socially killed Disney, which only found success by exploiting works that no longer held copyright. Etc.

    Those with power have lost all accountability, and all tools, especially AI, will be used against us if we do not cooperatively figure out how to fix the increasing power imbalance.

    The more power someone has, the harder the gavel should fall on them when they fuck the entire planet in whichever way.

    At this point, any new consumer friendly behaviour comes only to establish territory before hoarding and exploiting when enabled to do so.

    Amazon using deceptive design to influence general user behaviours should lead to billions and billions in fines until changed. Etc.

    Build local movements to cooperate at larger scale and fight back. If the general public is ranting about planned obsolescence and general monopolistic behaviours, maybe something could be affected before people are forced into violent desperation. People are too busy being mad at each other for some intentionally divisive narrative or another, and the general public just can’t give a fuck about affecting the people who actually dictate the shape of society.

    Also if you burn down all AI this is still true. But it’s easier to yell at technology than they system using it to further remove your autonomy.


  • will try to take it in good humour, but i love how i got compared to ai, adhd(AuDHD would be the real wombo combo here so you get points), and schizophrenic people.

    and i would hope i don’t confabulate half as much as an LLM.

    although an understanding of the modern situation does require an unfortunately theoretical take, while, unfortunately, there’s more noise, and conspiracy theories being socially reified than most people can remember. but i’d like to think i’m weighting this take via the best available expert consensus that i can find and source. biggest ‘correction’ i’d make is that i was beaten black and blue for waiting outside of the library, which was unrelated to the protest.

    if you do actually care, and can handle more than the internet’s usual 140 character tweet limit, here’s some elaboration.

    the ‘sycophancy into delusion effect’ i refer to can be seen widely reported on most news sites, where chatgpt and the like cause a feedback-loop into a psychotic break. this is one individual and machine, but a group that forgives the same things has the same sycophantic effect. predictive processing and the bayesian brain are leading theories in psychology that work well nested with other leading theories such as global workspace.

    that global workspace video is a very recent example with michael levin from tufts, who often works with friston’s free energy principle and active inference (included notes in wiki)

    friston has hundreds of thousands of citations, if you care about pedigree. i hope i do not poorly capture or inaccurately represent any of their ideas, but if you’d like to drink from the source, you have my full recommendation.

    that’s where the “saving energy” stuff comes from. while DKE might not perfectly and accurately explain the situation, i’m all for better ways to convey that eco-niche specific intelligence doesn’t always transfer, especially if it’s ‘overfit to a local minima.’ otherwise knowing you need high samples to gauge your intelligence in any particular niche is also related to the framework i’m describing. in the bio-world you have overspecialization, like pandas too fit to a specific environment, which may focus on skills that don’t transfer outside of that environment. there’s a lot more to gain from the full bayesian perspective, but there is a lot to be gained just by looking at how systems can successfully co-construct, and their possible failure states that are inevitable as systems grow apart into new niche environments.

    there’s actually an interplay between that ‘energy saving’ property and putting energy back out which can be used to explore the environment, build a more robust model, and survive greater environmental shifts. this is explained in active inference. good, but slightly old textbook on MITpress. lots of other online resources for the curious.

    i’m saying that meta-awareness of the failure states in these specific system dynamics could do much more general and robust good for society than being socially pressured into climbing the socio-economic hierarchy as hard as possible.

    there’s a term for an imagined AI going rogue due to being overfit to a single goal. this is called a ‘paperclip maximizer.’ i compare the current socio-economic system to that failure. you know, ‘capitalism number go up!’

    i don’t think any studies i’ve seen disagree with that take, but if there’s a relevant expert who’s got a strong weighting i’m unaware of, i’m always open to updating my weights.

    as for learning yourself into some information bubble, or how someone can hold ridiculous beliefs without the need to question them, such as grand confidence despite low evidence, is often by taking something you have low evidence about, and having high confidence. and then giving it a high weighting. funny enough, friston’s dysconnection hypothesis is about framing schizophrenia as precision weighting issues, but i don’t think they are the kind i have TY.

    mahault has a phd under friston, and her epistemic papers are essential IMO.

    so there you have it, the larger environment of my thoughts, largely focused around one of the most cited neuroscience experts of all time, and michael levin who i mentioned is doing some of the coolest current empirical results in modern biology.

    i tried, thank you if you got this far. if nothing else, please stay curious, but beware information silos that disable coms completely, or otherwise create barriers to properly comprehending the systems being represented. ‘nothing about us without us’ is important for a reason.

    otherwise, wish i could compress these complex topics into fewer words, but words are a lossy compression format.


  • Love this comment. If anyone knows anything about machine learning or brains, this resembles modal limitations in learning.

    A lot of our intelligence is shaped around our sensory experience, because we build tools for thinking via the tools we’ve already built, ever since baby motorbabbling to figure how our limbs work. Why Hellen Keller had such trouble learning, but once she got an interface she could engage with for communication, things took off.

    We always use different tools, but some people don’t see colour. This doesn’t mean they are stupid when they answer differently in describing a rainbow.

    Also why llms struggle with visual/physical concepts if the logic requires information that doesn’t translate through text well. Etc.

    Point being, on top of how shitty memorization is as the be all end all, learning and properly framing issues will have similar blindspots like not recognizing the anvil cloud.

    This is also why people in informational bubbles can confirm their own model from ‘learning’ over people’s lives experiences.

    Like most issues, it doesn’t mean throwing the baby out with the bathwater, but epistemic humility is important, and it is important not to ignore the possibility of blindspots, even when confidence is high.

    Always in context of the robustness of the framing around it, with the same rules applied at that level. Why “nothing about us without us” is important.

    But also we gotta stop people giving high confidence to high dissonance problems, and socializing it into law. We should be past the “mmr causes autism” debate by now, but I’m hearing it from the head of health in the USA.


  • I could see why you’d say that. Stress creates environments of basic survival, which kills cognitive thought. More immediate survival is more salient.

    That being said, if you have access to the internet, you have access to countless free educational tools.

    Too much privilege brings sycophantic bubbles of delusion, like billionaires.

    Having all the time and money also let’s you do a whole thing tank about how to ruin a country to fit your preferences. See the heritage foundation as prime example.

    That being said, while it is less easy for the poor, it’s still essential to attempt that open mind and learn, so you don’t get trapped by a socialized category error applied as fact.

    This is where we need predictive processing and the Bayesian brain to understand how beliefs are weighted and compared, and the failure states that might being.

    Basically, poor weighting or system communication leads to an over affirmation of something that should have been high uncertainty, if measured from other directions.

    Instead of seeing high cognitive dissonance as a sign to align low probability, it gets socialized into acceptance to save the energy of trying to worry about our deal with what, to that system, appears intractable.

    DKE is at least useful in framing how each expertise eco-niche is filled with complexity that doesn’t Transfer. This is why scientists stict to their expertise, where they have high dimensions of understanding, and low dissonance to uphold.

    This can be over-prioritized until no dissonance outside of microscopic niches that act more like data collection than science.

    Experts however can work together to find truths that diffuse dissonance generally, to continue building understanding.

    If the peasants could socialize that laziness was a lack of meta awareness of the greater dissonance diffusing web of shared expert consensus, instead of laziness being the act of not feeding the socio-economic hierarchy machine, which is famous for maximizing paperclips and crushing orphans.

    Pretty sure I got beaten black and blue waiting for library access. Had to protest to keep a library open when I’m gradeschool.

    So, growth mindset isn’t a privilege, but general access to affordances, pedigree, time, tools, social connections, etc, are all extra hurdles for growth mindset in impoverished places.

    If there’s no internet access at all, then that’s just a disabled system.

    Is not static with people, and Issue with growth mindset would just be vulnerability to learning yourself into some information bubble that intentionally cuts off communication, so that you can only use that group as a resource for building your world model, bringing you to where the closed brains go just to save energy, and keeping you there forever.

    Groups that are cool with making confident choices fueled by preference in high dissonance spaces. which basically acts like fertile soil for socializing strong cult beliefs and structures.

    They also use weird unconscious tools that keep them in the bubble. Listen to almost anyone that’s escaped a cult for good elaboration there. Our brains will do a lot to keep us from becoming a social pariah in our given environment we have grown into.