• 0 Posts
  • 11 Comments
Joined 1 month ago
cake
Cake day: October 9th, 2025

help-circle


  • Everyone seems to be tracking on the causes of similarity in training sets (and that’s the main reason), so I’ll offer a few other factors. System prompts use similar sections for post-training alignment. Once something has proven useful, some version of it ends up in every model’s system prompt.

    Another possibility is that there are features of the semantic space of language itself that act as attractors. They demonstrated and poorly named an ontological attractor state in the Claude model card that is commonly reported in other models.

    If the temperature of the model is set low, it is less likely to generate a nonsense response, but it also makes it less likely to come up with an interesting or original name. Models tend to be mid/low temp by default, though there’s some work being done on dynamic temperature.

    The tokenization process probably has some effect for cases like naming in particular, since common names like Jennifer are a single token, something like Anderson is 2 tokens, and a more unique name would need to combine more tokens in ways that are probably less likely.

    Quantization decreases lexical diversty, and is relatively uniform across models. Though not all models are quantized. Similarities in RLHF implementation probably also have an effect.

    And then there’s prompt variety. There may be enough similarity in the way in which a question/prompt is usually worded that the range of responses is restrained. Some models will give more interesting responses if the prompt barely makes sense or is in 31337 5P34K, a common method to get around alignment.




  • I ruminate on this from time to time. I’m not particularly well read on these things, but this is the closest I’ve managed to make sense of it: For whatever reason (and thar be many), the trumpite has embraced denigrating others as an acceptable way to soothe the ego. Unhealthy as it is, this kind of projection/displacement is by no means an uncommon trait in any demographic. Though it seems to be more common in groups that highly value hierarchy (as the Right typically does). On its own, utilizing this as a motivating force yields diminishing returns as shame or just boredom creep in. And so the aspiring demagogue hitches denigration to anger, goads projection with scapegoating, and ultimately harnesses hate to do evil.

    But this takes time. Trump didn’t skip all the way to invading cities and tearing down a third of the White House on day one. Trumpites have been lead gradually by these mechanisms and a host of complimentary social tactics to do more and more shameful things. And in doing, their egos are resiliently bridled to Trump’s. To turn against him now would mean facing feelings of abject humiliation and shame/guilt, and in some cases a complete reset of personal identity. Most people aren’t willing to do that until they encounter tragic and very personal consequences. So they persist in delusion, which comes naturally to parts of the religious demographics… but that’s a whole other can of worms.




  • The fact that workers with expense accounts still feel they’re getting paid so little that they deserve to commit fraud says something about that stratum of employee.

    Pretty much anyone who travels has to submit receipts. Most people who travel are not making bank. They’re the people who set up and stand at convention booths, sales staff support, assistants, videographers, etc. Also, most travel is a miserable ordeal. I’m not saying it’s okay to commit fraud, but let’s not equate the hourly employee “re-creating” his lost lunch receipt with a 6-figure income.



  • Eh, people said the exact same thing about Wikipedia in the early 2000’s. A group of randos on the internet is going to “crowd source” truth? Absurd! And the answer to that was always, “You can check the source to make sure it says what they say it says.” If you’re still checking Wikipedia sources, then you’re going to check the sources AI provides as well. All that changes about the process is how you get the list of primary sources. I don’t mind AI as a method of finding sources.

    The greater issue is that people rarely check primary sources. And even when they do, the general level of education needed to read and understand those sources is a somewhat high bar. And the even greater issue is that AI-generated half-truths are currently mucking up primary sources. Add to that intentional falsehoods from governments and corporations, and it already seems significantly more difficult to get to the real data on anything post-2020.