Anyone using DeepSeek as a service the same way proprietary LLMs like ChatGPT are used is missing the point. The game-changer isn’t that the Chinese company DeepSeek can compete with OpenAI and its ilk as an AI service provider—it’s that now any organization with a few million dollars to train and host their own model can compete with OpenAI.
I’d like to look into that, how can I train an existing model further?
I’m only playing around with ollama, but like to do a bit more - mostly just to fulfill my needs to understand things - but have no idea where to start
Python is not a problem
SW Dev is my job. Just never had real contact with AI before, besides playing around a bit.
Thank you very much for the link!!
Edit: thank you very much again, that was pretty much exactly what I was looking for.
Don’t know how I missed to checkout huggingface. Thought of it always just as a github for models and didn’t bother checking for docs…
But that’s a great intro with simple tools/tutorials to get a grip on it, thanks!
Anyone using DeepSeek as a service the same way proprietary LLMs like ChatGPT are used is missing the point. The game-changer isn’t that the Chinese company DeepSeek can compete with OpenAI and its ilk as an AI service provider—it’s that now any organization with a few million dollars to train and host their own model can compete with OpenAI.
Or open source groups can make a fully open repro of it: https://github.com/huggingface/open-r1
I’d like to look into that, how can I train an existing model further?
I’m only playing around with ollama, but like to do a bit more - mostly just to fulfill my needs to understand things - but have no idea where to start
You’re going to have to learn python.
Here’s a good overview: https://huggingface.co/docs/transformers/training
Python is not a problem
SW Dev is my job. Just never had real contact with AI before, besides playing around a bit.
Thank you very much for the link!!
Edit: thank you very much again, that was pretty much exactly what I was looking for.
Don’t know how I missed to checkout huggingface. Thought of it always just as a github for models and didn’t bother checking for docs…
But that’s a great intro with simple tools/tutorials to get a grip on it, thanks!
On-prem vs. Cloud, basically. On-prem just magically got cheaper.
Onprem has always been cheaper. Cloud compute was the most successful marketing campaign I can think of.