• 1 Post
  • 95 Comments
Joined 1 year ago
cake
Cake day: July 14th, 2023

help-circle
  • Code mods are great, maps and assets are in there but not officially, so compatibility going forward probably isn’t great for those. Full modding support is being worked on and is one of their highest priorities, so I’m not surprised there wasn’t much discussion about it. Asset mod support is “before summer” so they’ve got another month according to their last statement on it. PDX Mods has some bugs but overall it’s actually pretty slick and functional, and they’ve made a few highly requested improvements to it already.



  • If you read the full article, it seems as if the Saudi religious establishment was infiltrated by Egyptian extremists fleeing a crackdown on the Muslim Brotherhood following the assassination of Sadat. Their ideology meshed with Wahhabism and Bin Laden’s religious vendetta against the United States. The Saudi state apparatus did not have effective oversight over the religious establishment and so this all happened under the House of Saud’s nose. The countries in red are (at the time) places with either US puppet regimes or some form of Arab Revolt descended, nominally secular/socialist regimes. The religious extremists pushing Islamic rule operated in these countries under various militias and terrorist groups, notably Al Qaeda, backed by the newly radicalized Saudi Wahhabi establishment, and of course, Iran.

    From that perspective, the US was waging war against militias and terrorist groups with roots and support in Saudi Arabia, but the House of Saud was not considered to be complicit. The article goes on to say…

    Astonishingly, the attacks of 9/11 had little effect on the Saudi approach to religious extremism, as diplomats and intelligence officials have attested. What finally changed royal minds was the experience of suffering an attack on Saudi soil. In May 2003, gunmen and suicide bombers struck three residential compounds in Riyadh, killing 39 people. The authorities attributed the attacks to al-Qaeda, and cooperation with the U.S. improved quickly and dramatically.

    Interesting stuff, to be sure.



  • In the US, there are positive and negative stereotypes, too. German efficiency and Japanese perfectionism and perseverance are among them. Jewish intelligence and commitment to education, too. These things have a basis in reality, of course, but they shouldn’t be mistaken for reality itself. It seems to me these things appearing in your textbooks were probably attempts by your own government to get its people to emulate what it sees as positive traits in other cultures, rather than an attempt by foreign adversaries to paint Chinese people as inferior. Of course, when the message was a little too unclear or negative as in the “toxic textbooks” incident, your government deflected blame.






  • Let’s just take NYT for example. Subscription costs $325/year. Why would I ever pay that much? It’s not 1954. I’m not sitting down with my morning coffee and reading the damn thing front to back. I’m reading maybe one article a week from 15 different sources. Am I supposed to pay $5000/year just to cover my bases?

    As with everything else in [CURRENT YEAR] the value proposition is so absurdly out of step with reality that fixing it basically relies on rolling out the guillotines.








  • Bought a Win 11 laptop recently (it’s now a linux device), and on first boot up there was no way to gracefully decline using a Microsoft account to sign in. Luckily, I was in a hotel and couldn’t connect to the wifi without going through a login page, so the lack of internet connection allowed me to set up a local account. In any case, if you’re forced to log in with an MS account, OneDrive starts syncing right away. You can disable it, but maybe not before it’s already done some damage.




  • I guess I’m wondering if there’s some way to bake the contextual understanding into the model instead of keeping it all in vram. Like if you’re talking to a person and you refer to something that happened a year ago, you might have to provide a little context and it might take them a minute, but eventually, they’ll usually remember. Same with AI, you could say, “hey remember when we talked about [x]?” and then it would recontextualize by bringing that conversation back into vram.

    Seems like more or less what people do with Stable Diffusion by training custom models, or LORAs, or embeddings. It would just be interesting if it was a more automatic process as part of interacting with the AI - the model is always being updated with information about your preferences instead of having to be told explicitly.

    But mostly it was just a joke.