I’m forever blowing bubbles

by liberal japonicus

A large amount of my reading and listening has been concerned with bubbles, specifically the one linked to so called "AI". Some fun etymology before I get into the actual topic. Bubble seems to be from Dutch or German, but I guess is onomatopoeic. About bubbles, the Japan bubble got me here, thinking that I would be king of the world if I made it here (whoops) so bubble also raises up the Yiddish word bubbeleh, a term of endearment. So seeing another bubble coming up is almost like finding a nostalgic candy I used to eat as a kid. 

In Japanese, the economic phenomenon uses the katakana word バブル (baburu), but Japan had experienced speculative booms and busts before Meiji. Japan had the first futures market, the Dojima Rice Exchange, so it is interesting to me that Japanese borrows a word for something that must have been identified earlier. Some googling suggests that the phrase used was 相場狂乱 (sōba kyōran), which translates to 'price frenzy/madness. The Japanese kanji for bubble is 泡 (awa), but rather than referring to a single bubble, it tends to refer to foam i.e. a group of bubbles. Which is interesting, because in Greek, foam is associated with Aphrodite (the name means she who emerges from foam) and, like Japanese, doesn't transfer to a single bubble. So a bubble not only refers to a singular phenomenon, I also take it to imply that it is something f**ked up that shouldn't normally exist in nature. 

Anyway, been doing stuff with AI, specifically Gemini and Notebooklm, to get ready for classes. Going with these as opposed to any others because they are folded into the students' Google accounts. But the title will probably get into what I thought we might talk about, which is whether we are in bubble territory or not. Some observations to kick this thing off. Unfortunately, I can't reference these, cause almost all of them have been snagged from various podcasts/youtube videos, but the first thing I want to say is that using the term AI for LLM was marketing genius, on par with getting people to call a paper tissue Kleenex or a copy a Xerox. Perhaps more, because an LLM is not AI, except that naming makes it so. Here's a list of other questions, gleaned from various articles, podcasts and youtube videos that have all jumbled together.

-Is there an AI bubble? I would say yes, the disappointment with the release of ChatGPT 5 suggests that LLMs are hitting a wall with scaling, and having Altman saying that there was over-excitement in AI as proof. It's amazing that the expenditures on AI may be propping up the economy in the face of Trump's policies on trade and the cratering of tourism to the US.

-Are these tools game changing? That is also yes. One podcast pointed out that the internet basically nuked travel agents, and to that example, you could add bookstores, record stores and film processors and it seems evident that the same kind of process will be occurring for this.

-Have LLMs hit a wall? I'd say yes to that too, though the phrase 'fake it til you make it' does come to mind. Frex, ChatGPT (and others) apparently can take in enough chess to play, but will, when pressed, play illegal moves. This can be explained by them vacuuming up games in chess notation, but not being taught the rules, so they can generalize an apparent ability to play. But the minimal gains shown in chatgpt 5 are telling, and the increasing amount of massaging that is needed is getting to be obvious. 

For a concentrated dose of de-hype, I recommend Gary Marcus, though because he's been saying the king has not clothes for a while (and shit on for it), he's a bit on the strident side. 

Anyway, an open thread for all your LLM needs

17 thoughts on “I’m forever blowing bubbles”

  1. using the term AI for LLM was marketing genius
    The phenomenon is much more widespread than just slapping the “AI” label on LLMs. Almost anything newsworthy that’s done by a computer these days is hyped as “AI.”
    See, for example, this article about Delta’s pricing. Someone more knowledgeable than I (which is probably almost everyone here) can correct me if I’m wrong, but Delta’s personalized pricing seems to require a big database of personal data that shouldn’t be public in the first place (a whole topic in its own right) and an algorithm for analyzing it. In other words, a bog standard computer program.
    Another example: a friend of mine recently checked out the conversational capabilities at one of the well known language teaching sites. (I don’t know which one.) He labeled the conversation as “AI”-generated, which may in fact be accurate, but it reminded me of Eliza, which was written ~60 years ago.
    And then there’s this, about which I got nothin’.
    The mention of travel agents and bookstores is a good reminder to skeptical me that “AI” is going to have widespread disruptive effects even if it isn’t remotely what it’s hyped to be.

  2. Bubble or not, the current LLMs and other types of “AI” will have a massive impact even if development hits a wall.
    I’m currently using Grok for most things because it seems to have more horsepower behind it than the free versions of other LLMs.
    The wordsmiths among you may find much to criticize, but this review of the TV drama series, Veronica Mars, by Grok seems very good to me.
    Veronica Mars: Noir, Trauma, and Justice

  3. It’s amazing that the expenditures on AI may be propping up the economy in the face of Trump’s policies on trade and the cratering of tourism to the US.
    It somewhat depends on where you are. I’ve been seeing stories of people using local electric grids being overwhelmed because someone set up an AI (LLM) data center in the area. They are incredible users of electricity. Can be a real issue if your power goes out as a result of the extra load.

  4. The old Eliza chatbot was generally harmless.
    Wasn’t there another, called something like ‘Racktor’, that simulated a paranoid schizophrenic?
    Seems like Trump could be replaced with that one, with little observable difference.

  5. The program was ‘Racter’, short for ‘raconteur’ (story teller) but the system allowed only 6 characters for file names, and dates from 1984.
    The difference to His Orangeness would be huge since the program at least made no errors in grammar or spelling. And at least the texts published by Chamberlain were far more entertaining than the utterings (in text and speech) of Jabbabonk the Orange.
    https://en.wikipedia.org/wiki/Racter

  6. Seems like Trump could be replaced with that one, with little observable difference.
    Or a monkey that flings tariffs instead of shit.

  7. Seems like Trump could be replaced with that one, with little observable difference.
    Au contraire
    , it would definitely be easier on the eyes. “Observable” in the literal meaning of the word.

  8. ij – sounds like you are thinking through some of the issues that Christopher Newfield discusses in his Critical AI article “How to Make ‘AI’ Intelligent; or, The Question of Epistemic Equality.” [https://doi.org/10.1215/2834703X-10734076%5D I don’t see that the article is open access, so I’ll excerpt a chunk of the intro here to give y’all a sense of what Newfield is arguing…
    In this article I will not provide a historical account of how AI research in its various iterations has alleged rather than specified the intelligence it aims to simulate. Instead, I will first suggest why it is so hard for most technologists and the officials they influence to care about rigorous definitions of intelligence before they attribute it to software. I will then analyze one philosopher’s rigorous definition for its implications for current debates about AI.
    Some of the problem is the size and wealth of the corporate platforms that have dominated the internet and its opportunities for data accumulation and that now want to dominate “AI.” But there is also a deeper and more difficult cultural problem, and that is the highly restricted role of culture itself, or the role of cultural analysis of technology. There have been frequent periods when technologists are able, intentionally or not, to keep cultural analysis from having coauthorship of the meaning, operations, and effects of the technology or of narratives of its future. “The Age of AI” is one of those periods when practitioners who bring culturally grounded skepticism to technological development are less likely to be treated as equal partners than to find themselves unemployed, as did computer scientist Timnit Gebru in 2020 (Simonite 2020a, 2020b).
    This is happening in spite of the fact that the long history of asking “What is intelligence?” also belongs to historical and cultural disciplines—philosophy obviously, but also to feminist studies, which radicalized the context dependence of knowledge in standpoint theory, and ethnic studies, which demonstrated the role of ascribed race in structuring epistemological frameworks, just to name two of the many domains of philosophically informed cultural research. Of course major contributions to this question have been made by scientists and technologists, but these have been boundary crossers, traveling back and forth between cultural and technological disciplines and bringing their procedures and findings together. There is much discourse from practitioners affiliated with AI-labeled technology about the technology’s power to benefit all humanity. This is not the same as a full discussion among epistemic equals about whether this assertion is actually true, whether our diverse societies want it to be, and what would make it true in ways that diverse societies might want.
    This is a deep, complicated, and massively multilateral conversation that will take years or decades. It requires much better public education processes than what we see today in the mainstream media.1 It means continuous travel of informed people among multiple disciplines and synthesis of disparate methods and their results. This is not happening. AI discourse is largely a question of when, not if, and it assumes that technologies will be pushed out to the consuming masses by large corporations and the start-ups they fund on a schedule that they determine. A (problematic) call for a pause in AI research presumed that the revolution of superior machine intelligence is here and that proper management ensures an unspecified flourishing future (Future of Life Institute 2023).2 AI discourse often functions as a manifest destiny about which great minds are said to agree.
    At the same time, critics have identified a set of operational flaws in the ML and deep learning systems now discussed under the “AI” banner. Four of the most discussed are social biases, particularly racism, that become part of both the model and its use; opacity, such that users cannot assess how results were generated; coercion, in that architectures, datasets, algorithms, and the like are controlled by designers and platforms rather than users; and privacy violations, which result from combinations of bias, opacity, and coercion focused on the surveillance, accumulation, and monetization of data.
    Readers of Critical AI are among those increasingly focused on a fifth operational flaw: much ML research takes place in companies like Google, in which managers have authority over the publication of research results. Famous cases like the one I mentioned above, the firing of Google ethics researcher Timnit Gebru (and her co-lead Margaret Mitchell), suggest that much or most AI research is happening in the absence of academic freedom, which puts researchers at risk while also distorting research results by allowing the suppression of findings that don’t fit a rollout narrative or corporate image. Corporate manipulation of research results is a known issue thanks to the automotive, tobacco, chemicals, and fossil fuel industries, among others.
    Then there is a sixth issue that I’m considering here—the question of whether “AI” is intelligent in the first place. And there is the related question of why this sixth question is not central to public AI debates.
    Reflecting on the work of two authors can help us address these underexamined questions. The first is C. P. Snow and his famous meditations on the divide between “Two Cultures” (which Snow described as scientific vs. literary outlooks but which I will discuss in terms of technological vs. cultural knowledge). The second is programmer and philosopher Brian Cantwell Smith’s recent analysis in The Promise of Artificial Intelligence (2019) of two kinds of intelligence: reckoning and judgment. Smith sheds light on the mentality that the editors of this special issue identify as data positivism, but which Smith’s notion of “reckoning” helps me talk about more explicitly as computational intelligence of a certain kind. Snow helps explain why culture-based understandings of intelligence are not part of the current debate, and Smith shows what can happen when they are. Although Snow did not intend to, he helped take humanities disciplines out of the future-defining process for several generations. Smith offers us ways of putting them back in.

    In the body of the article Newfield fleshes out these points a lot and settles into a discussion of intelligence as not just a matter of number crunching and pattern recognition (‘reckoning’) but also of what Smith (mentioned above) calls ‘registration’ (a situated perception of the world and the reckoner’s place in that system) and also of judgment (which I will gloss as a commitment to reconciling reckoning and registration in order to test and negotiate a shared understanding that fits both the data and the human relationships that are entangled/implicated in that data, or as Newman interprets Smith: “‘The system (knower) must be committed to the known, for starters. That is part of the deference’ in which one defers to the object in order to know it. But there’s a further matter, in which the knower must be ‘committed to tracking things down, going to bat for what is right,’ and feeling in some deep way existentially ‘bound by the objects’ (93). Intelligence, Smith is saying, depends on an underlying awareness of existence, both one’s own existence and the existence of the world. For Smith, epistemology is prior to ontology, and we can add that a feeling of existence is prior to them both—and a fundamental precondition of intelligence.).”
    In terms of Charles’ Veronica Mars example, Grok has “reckoned” what others have communicated about the show, but in assembling its commentary it cannot exercise judgment because it does not understand the cultural systems in which those communications gain or lose their significance, nor situate itself in relation to the various parties involved in the communication.
    It has grokked nothing, it has only parsed and systematized according to its own training algorithms.
    The Veronica Mars output is a “good” summary in that it assembles together a lot of information and represents it in a way that appears faithful, but it doesn’t rise to the level of any critical insight, and it does not provide any transparency for the sources of its information or the reasoning behind its selection. It feels like an intellectual cul-de-sac to me.

  9. …, and it does not provide any transparency for the sources of its information…
    It usually provides links to its sources. I’ll have to add a source links requirement to the prompt.

  10. CharlesWT – It usually provides links to its sources. I’ll have to add a source links requirement to the prompt.
    Links to sources would be helpful, but the deeper issue of transparency involves the selection criteria that results in those sources being included. It’s one of the questions I routinely ask my lower division undergraduate students when it comes to their own papers: “what purpose does this citation serve in developing an understanding of the critical perspective from which you are writing?”
    Newfield addresses this deeper sense of transparency in the article intro where he writes: “At the same time, critics have identified a set of operational flaws in the ML and deep learning systems now discussed under the “AI” banner. Four of the most discussed are social biases, particularly racism, that become part of both the model and its use; opacity, such that users cannot assess how results were generated; coercion, in that architectures, datasets, algorithms, and the like are controlled by designers and platforms rather than users; and privacy violations, which result from combinations of bias, opacity, and coercion focused on the surveillance, accumulation, and monetization of data. What I’m pointing to here falls under the second and third operational flaws. We don’t know why the LLM chose these particular points to amplify. They are as opaque to us as the proprietary systems by which search results get ordered on search engines.
    The lack of epistemological and methodological awareness are a deep problem, and these are the reasons why I scoff at Altman’s comparison of the latest iteration of ChatGPT as being like interacting with an expert with a Ph.D.. The lack of these deeper levels of awareness are more a marker of someone much earlier in their intellectual development.

  11. Thanks nous, research hub lets me request a full text, so I’ll try that.
    I’ve made a post out of GftNC’s depressing news.

  12. I’ve been seeing stories of people using local electric grids being overwhelmed because someone set up an AI (LLM) data center in the area. They are incredible users of electricity.
    “HighYield x SemiAnalysis deep-dive into AI Datacenters, Gigawatt Megaclusters and the Hyperscaler race to AGI. How AI Datacenters Eat the World.”
    How AI Datacenters Eat the World

Comments are closed.