• 2 Posts
  • 15 Comments
Joined 4 months ago
cake
Cake day: December 9th, 2024

help-circle

  • If by more learning you mean learning

    ollama run deepseek-r1:7b

    Then yeah, it’s a pretty steep curve!

    If you’re a developer then you can also search “$MyFavDevEnv use local ai ollama” to find guides on setting up. I’m using Continue extension for VS Codium (or Code) but there’s easy to use modules for Vim and Emacs and probably everything else as well.

    The main problem is leveling your expectations. The full Deepseek is a 671b (that’s billions of parameters) and the model weights (the thing you download when you pull an AI) are 404GB in size. You need so much RAM available to run one of those.

    They make distilled models though, which are much smaller but still useful. The 14b is 9GB and runs fine with only 16GB of ram. They obviously aren’t as impressive as the cloud hosted big versions though.








  • I’m noticing a trend of scientific-sounding announcements about physics results that turn out to be theoretical explorations of simulations. The whole “false vacuum” idea isn’t really even a hypothesis, just a what-if. We have no indication that the ground vacuum state isn’t the lowest energy configuration. I think people just find a non-zero minimum unintuitive.

    Anyway, the key figure in all these theoretical simulation articles is the multi-billion dollar quantum super computers running these simulations. Wouldn’t it be funny if tech investors with a lot of money staked on quantum devices pushed for low-quality science that required their machines to be done, thus expanding the market and value of their otherwise pointless supercomputers? This article ends on a very optimistic “these computers have so many uses in cryptography and science” which seems a little out of place when discussing physics results.