I spin up a lot of Docker containers with large data sets locally.
Developer here. Completely depends on your workflow.
I went base model and the only thing I regret is not getting more RAM.
Speeds have been phenomenal when there binaries are native. Speeds have been good when the binaries are running through Rosetta.
The specs you’re wavering between are extremely workflow specific. You know if your workflow requires the 16 extra GPU cores. You know if your workflow requires another 64 GB of RAM.
Raising the standard enables new uses of technology.
Serious. I installed VSCodium today.
Use two providers on different networks. They can fill in the gaps for each other.
Chris from Mr. Beast is non-binary.
Link from below:
https://edition.cnn.com/2023/04/15/entertainment/mr-beast-transphobia-chris-tyson-trnd/index.html
Yet the same people still expect meteorologists to use the same science to predict the weather for them.
I highly recommend storing your DB and pictrs directories on an SSD volume.
I’m running on my NAS.
So far so good on my little one user instance as well.
Exactly what you said. It has always been about control.