Update blog post with pics

This commit is contained in:
Jake Runyan 2025-01-23 09:50:58 -08:00
parent 81ac3e7a86
commit 6956bbc25b
3 changed files with 47 additions and 0 deletions

View File

@ -0,0 +1,47 @@
---
slug: pure-rl-with-deepseek
title: Pure RL with DeepSeek
authors: [jrunyan]
tags: [ai, workflow]
---
# Pure RL with DeepSeek
So apparently pure reinforcement learning is the move. The new DeepSeek models out of China throw modern LLM training
techniques out in favor of just purely using RL over more time and over more data to produce better models than the more bespoke methods.
# Open Source Stacks Rock!
I've experienced it once before with [ComfyUI](https://jake.runyan.dev/blog/sdxl-pipeline), but it seems like with AI the open source
community has really been putting in the work so those like me looking to start some passion projects can quicken their development cycles.
[OpenWebUI](https://github.com/ml-explore/OpenWebUI) is a great frontend UI for interacting with models. Some of their docker containers
come bundled with Ollama, which means the setup is literally as simple as building a super standard docker container with it.
# Giving it a Shot
## DeepSeek
This is the new LLM model out of China that's been said to have pretty good code gen abilities, and the main reason I started looking
at setting up this stack for myself. I've done previous attempts at self-hosting coding assistants, but found that the juice wasn't
worth the squeeze, as they say.
For this experiment I was looking at the 1.5b and 8b models for `deepseek-r1`, though the coding assistant model `deepseek-coder` is also available.
## My Experience
Man, it was pretty good. I've got a smaller GPU on PWS so I was limited to running the 8b model, but responses were good. I noticed
the best response times on the 1.5b model, and for some easier tasks the correctness between the two was hard to discern.
I particularly liked the features of OpenWebUI to allow for web search, which from initial testing seemed to find good results to build
context of the response with.
![img alt](./web-search.png)
Some responses with web search are a little on the nose with the smaller models, for example a search about me gives results that are pulled
verbatim from my github profile and websites.
![img alt](./jake.png)
I'm still actively using chatgpt, claude, and others for coding work, but as local LLMs improve you can bet I'll be keeping up to date with this stack.
## Resources
[OpenWebUI](https://github.com/ml-explore/OpenWebUI)
[Deepseek with Ollama](https://ollama.ai/library/deepseek-coder)
Thank you to [DWS](https://dws.rip) for collaboration.

Binary file not shown.

After

Width:  |  Height:  |  Size: 108 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 48 KiB