mirror of
https://github.com/runyanjake/jake.runyan.dev.git
synced 2025-10-04 05:47:30 -07:00
Update blog post with pics
This commit is contained in:
parent
81ac3e7a86
commit
6956bbc25b
47
website/blog/2025-01-22-pure-rl-with-deepseek/index.mdx
Normal file
47
website/blog/2025-01-22-pure-rl-with-deepseek/index.mdx
Normal file
@ -0,0 +1,47 @@
|
||||
---
|
||||
slug: pure-rl-with-deepseek
|
||||
title: Pure RL with DeepSeek
|
||||
authors: [jrunyan]
|
||||
tags: [ai, workflow]
|
||||
---
|
||||
|
||||
# Pure RL with DeepSeek
|
||||
So apparently pure reinforcement learning is the move. The new DeepSeek models out of China throw modern LLM training
|
||||
techniques out in favor of just purely using RL over more time and over more data to produce better models than the more bespoke methods.
|
||||
|
||||
# Open Source Stacks Rock!
|
||||
I've experienced it once before with [ComfyUI](https://jake.runyan.dev/blog/sdxl-pipeline), but it seems like with AI the open source
|
||||
community has really been putting in the work so those like me looking to start some passion projects can quicken their development cycles.
|
||||
|
||||
[OpenWebUI](https://github.com/ml-explore/OpenWebUI) is a great frontend UI for interacting with models. Some of their docker containers
|
||||
come bundled with Ollama, which means the setup is literally as simple as building a super standard docker container with it.
|
||||
|
||||
# Giving it a Shot
|
||||
|
||||
## DeepSeek
|
||||
This is the new LLM model out of China that's been said to have pretty good code gen abilities, and the main reason I started looking
|
||||
at setting up this stack for myself. I've done previous attempts at self-hosting coding assistants, but found that the juice wasn't
|
||||
worth the squeeze, as they say.
|
||||
|
||||
For this experiment I was looking at the 1.5b and 8b models for `deepseek-r1`, though the coding assistant model `deepseek-coder` is also available.
|
||||
|
||||
## My Experience
|
||||
Man, it was pretty good. I've got a smaller GPU on PWS so I was limited to running the 8b model, but responses were good. I noticed
|
||||
the best response times on the 1.5b model, and for some easier tasks the correctness between the two was hard to discern.
|
||||
|
||||
I particularly liked the features of OpenWebUI to allow for web search, which from initial testing seemed to find good results to build
|
||||
context of the response with.
|
||||
|
||||

|
||||
|
||||
Some responses with web search are a little on the nose with the smaller models, for example a search about me gives results that are pulled
|
||||
verbatim from my github profile and websites.
|
||||
|
||||

|
||||
|
||||
I'm still actively using chatgpt, claude, and others for coding work, but as local LLMs improve you can bet I'll be keeping up to date with this stack.
|
||||
|
||||
## Resources
|
||||
[OpenWebUI](https://github.com/ml-explore/OpenWebUI)
|
||||
[Deepseek with Ollama](https://ollama.ai/library/deepseek-coder)
|
||||
Thank you to [DWS](https://dws.rip) for collaboration.
|
BIN
website/blog/2025-01-22-pure-rl-with-deepseek/jake.png
Normal file
BIN
website/blog/2025-01-22-pure-rl-with-deepseek/jake.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 108 KiB |
BIN
website/blog/2025-01-22-pure-rl-with-deepseek/web-search.png
Normal file
BIN
website/blog/2025-01-22-pure-rl-with-deepseek/web-search.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 48 KiB |
Loading…
x
Reference in New Issue
Block a user