• 0 Posts
  • 71 Comments
Joined 2 years ago
cake
Cake day: June 14th, 2023

help-circle
  • I don’t understand how people can look at the insane progress gpt has made in the last 3 years and just arbitrarily decide that this is its maximum capability.

    So this is not entirely arbitrary, and probably part of it is also that they’re not just looking at the progress, but also at systemic issues.

    For example we know that larger models with more training material are more powerful. That’s probably the biggest contributing factor to the insane pace at which they’ve developed. But we’re also at a point where AI companies are saying they are running out of data. The models we have now are already trained on basically the entire open internet and a lot of non-public data too. Therefore we can’t expect their capabilities to scale with more data unless we find ways to get humans to generate more data. At the same time the quality of data on the open internet decreases because more of it is generated by AI.

    On the other hand, making them larger also has physical requirements, most of all power. We are already at a point where AI companies are buying nuclear power plants for their data centers. So scaling in this way is close to the limit too. Building new nuclear power plants takes ages.

    Another different thing is that LLMs can’t learn. They don’t have to be able to learn to be useful, obviously we can use the current ones just fine at least for some tasks. But nonetheless this is something that limits the progress that’s possible for them.

    And then there is the entire AI bubble thing. The economical side of things, where we have an entire circular economy based on the idea that companies like OpenAI can spend billions on data centers. But they are losing money. Pretty much none of the AI companies are profitable other than the ones that only provide the infrastructure. Right now investors are scared enough to miss out on AGI to continue investing but if they stopped, it would be over.

    And all this is super fragile. The current big players are all using the same approach. If one company makes that next step and finds a better approach than transformer LLMs, the others are toast. Or if some Chinese company makes a breakthrough with energy usage again. Or if there is a hardware breakthrough and the incentive to pay for hosted LLMs goes away. Basically even progress can pop the bubble because if we can all run AI that does a good enough job at home then the AI companies will never hit their revenue targets. And then the investment stops and companies that bleed billions every quarter without investors backing them can die very quickly.

    Personally I don’t think they will stop becoming better right now. Even if they do stop, I’m not convinced we understand them well enough to be unable to improve the ways in which we use them a bit more. But when people say that this is the peak, they’re looking at the bigger picture. They say that LLMs can’t get closer to human intelligence because fundamentally, we don’t have a way to make them learn, they say that the development model is not sustainable, and other reasons like that.








  • On Mac:

    If you want an icon you can double click on your desktop, you can put you command in a file with the extension “.command” and mark it as executable. Double clicking it will run the content as a shell script in Terminal.

    If you want something that can be put into the Dock, use the Script Editor application that comes with macOS to create a new AppleScript script. Type do shell script "<firefox command here>" then find Export in the menu. Instead of Script, choose export to Application and check Run Only. This will give you an application you can put in the Dock.

    If you want to use Shortcuts, you can use the Run Shell Script action in Shortcuts too.

    Finally, if you want something that opens multiple firefoxes at once, chain multiple firefox invocations together on one line separated by an ampersand. There is an option you have to use (–new-instance I think?) to make Firefox actually start a complete new instance.


  • That’s funny because I grew up with math teachers constantly telling us that we shouldn’t trust them.

    Normal calculators that don’t have arbitrary precision have all the same problems you get when you use floating point types in a programming language. E.g. 0.1+0.2==0.3 evaluates to false in many languages. Or how adding very small numbers to very large numbers might result in the larger number as is.

    If you’ve only used CAS calculators or similar you might not have seen these too since those often do arbitrary precision arithmetics, but the vast majority of calculators is not like that. They might have more precision than a 32 bit float though.



  • What bothers me the most is the amount of tech debt it adds by using outdated approaches.

    For example, recently I used AI to create some python scripts that use polars and altair to parse some data and draw charts. It kept insisting to bring in pandas so it could convert the polars dataframes to pandas dataframes just for passing them to altair. When I told if that altair can use polars dataframes directly, that helped, but two or three prompts later it would try to solve problems by adding the conversion again.

    This makes sense too, because the training material, on average, is probably older than the change that enabled altair to use polars dataframes directly. And a lot of code out there just only uses pandas in the first place.

    The result is that in all these cases, someone who doesn’t know this would probably be impressed that the scripts worked, and just not notice the extra tech debt from that unnecessary dependency on pandas.

    It sounds like it’s not a big deal, but these things add up and eventually, our AI enhanced code bases will be full of additional dependencies, deprecated APIs, unnecessarily verbose or complicated code, etc.

    I feel like this is one aspect that gets overlooked a bit when we talk about productivity gains. We don’t necessarily immediately realize how much of that extra LoC/time goes into outdated code and old fashioned verbosity. But it will eventually come back to bite us.


  • I have to do a bunch of relatively unsurmountable steps to do what should’ve taken half a minute. Like screenshot the profile and scrape the text with iOS Photos text recognition.

    The iOS workaround isn’t quite as unsurmountable as you don’t have to go through the Photos app at all. You can enter text selection mode directly from the screenshot without even saving it or leaving the app you’re in. And since iOS will look up any word you can select in the system dictionary and also translate any text you can select, you can do these things right there too.

    That said I did once make a shortcut that lets me triple tap the back of my phone to pop up a text version of everything on screen that the iOS OCR detects. Not sure what I did that for though, I don’t really use it.







  • I tried it and it’s way off for me because it gives too much weight to submitted posts. I don’t have very many submissions so even when I selected recent only, it focused on one guide post for a game I wrote many years ago and made the profile 80% about that. But I guess that’s a problem at some point before the LLM is involved. There are some other similarly non-LLM problems too like making the most used terms section list almost only subreddit names.

    When I limited it to recent comments only it did a better job. It even listed “Humanity’s general incompetence” as the fifth of my “top 3” topics.