aboutsummaryrefslogtreecommitdiff
path: root/content/blog/2024-01-13-local-llm.md
diff options
context:
space:
mode:
authorChristian Cleberg <hello@cleberg.net>2024-07-02 23:24:15 -0500
committerChristian Cleberg <hello@cleberg.net>2024-07-02 23:24:15 -0500
commit89e6c0f9c8edd029aeeb553dcb11b62c878d11b1 (patch)
tree55a4fd7d3de25de21fe296c8353642bae1e4579b /content/blog/2024-01-13-local-llm.md
parentc5a2b7970f2187456dd094105e38b1d5482cb90f (diff)
downloadcleberg.net-89e6c0f9c8edd029aeeb553dcb11b62c878d11b1.tar.gz
cleberg.net-89e6c0f9c8edd029aeeb553dcb11b62c878d11b1.tar.bz2
cleberg.net-89e6c0f9c8edd029aeeb553dcb11b62c878d11b1.zip
fix 2024 blog posts for missing images
Diffstat (limited to 'content/blog/2024-01-13-local-llm.md')
-rw-r--r--content/blog/2024-01-13-local-llm.md13
1 files changed, 0 insertions, 13 deletions
diff --git a/content/blog/2024-01-13-local-llm.md b/content/blog/2024-01-13-local-llm.md
index ede130e..b6e1f86 100644
--- a/content/blog/2024-01-13-local-llm.md
+++ b/content/blog/2024-01-13-local-llm.md
@@ -50,16 +50,12 @@ After running the app, the app will ask you to open a terminal and run the
default `llama2` model, which will open an interactive chat session in the
terminal. You can startfully using the application at this point.
-![Ollama](https://img.cleberg.net/blog/20240113-local-llm/ollama.png)
-
If you don't want to use the default `llama2` model, you can download and run
additional models found on the [Models](https://ollama.ai/library) page.
To see the information for the currently-used model, you can run the `/show
info` command in the chat.
-![Model Info](https://img.cleberg.net/blog/20240113-local-llm/ollama_info.png)
-
## Community Integrations
I highly recommend browsing the [Community
@@ -68,9 +64,6 @@ section of the project to see how you would prefer to extend Ollama beyond a
simple command-line interface. There are options for APIs, browser UIs, advanced
terminal configurations, and more.
-![Ollama
-SwiftUI](https://img.cleberg.net/blog/20240113-local-llm/ollama-swiftui.png)
-
# iOS
While there are a handful of decent macOS options, it was quite difficult to
@@ -93,12 +86,6 @@ Once you have a file downloaded, you simply create a new chat and select the
downloaded model file and ensure the inference matches the requirement in the
`models.md` file.
-See below for a test of the ORCA Mini v3 model:
-
-| Chat List | Chat |
-| ----------------------------------------------------------------------- | ----------------------------------------------------------------- |
-| ![](https://img.cleberg.net/blog/20240113-local-llm/llm_farm_chats.png) | ![](https://img.cleberg.net/blog/20240113-local-llm/llm_farm.png) |
-
[Enchanted](https://github.com/AugustDev/enchanted) is also an iOS for private
AI models, but it requires a public-facing Ollama API, which did not meet my "on
device requirement." Nonetheless, it's an interesting looking app and I will