I compared the Ollama token/s for multiple GPUs for a handful of large language models: Older Nvidia GPUs are still performing pretty well. https://blog.stefandroid.com/2025/06/02/benchmark-llm-performance-nvidia-gpus.html?utm_source=mastodon
I compared the Ollama token/s for multiple GPUs for a handful of large language models: Older Nvidia GPUs are still performing pretty well. https://blog.stefandroid.com/2025/06/02/benchmark-llm-performance-nvidia-gpus.html?utm_source=mastodon
#Ollama v0.8 segna un importante passo avanti nell’interazione con i #LLM.
Dettagli e test: https://www.linkedin.com/posts/alessiopomaro_ollama-llm-qwen-activity-7334823072337653760-B6-G
___ 𝗦𝗲 𝘃𝘂𝗼𝗶 𝗿𝗶𝗺𝗮𝗻𝗲𝗿𝗲 𝗮𝗴𝗴𝗶𝗼𝗿𝗻𝗮𝘁𝗼/𝗮 𝘀𝘂 𝗾𝘂𝗲𝘀𝘁𝗲 𝘁𝗲𝗺𝗮𝘁𝗶𝗰𝗵𝗲, 𝗶𝘀𝗰𝗿𝗶𝘃𝗶𝘁𝗶 𝗮𝗹𝗹𝗮 𝗺𝗶𝗮 𝗻𝗲𝘄𝘀𝗹𝗲𝘁𝘁𝗲𝗿: https://bit.ly/newsletter-alessiopomaro
Say hello to kuzco, an LLM-powered computer vision assistant package in #RStats!
Frank Hull's kuzco leverages #Ollama to bring computer vision capabilities into your R workflow.
Learn about classification, image recognition, and sentiment analysis in kuzco: https://posit.co/blog/kuzco-computer-vision-with-llms-in-r/
New Ollama Release!
Version: v0.9.0
Release Notes:

## New models
- [DeepSeek-R1-0528](https://ollama.com/library/deepseek-r1:8b): A new version of DeepSeek-R1 with significantly improved depth of reasoning and inference capabilities.
```
ollama pull deepseek-r1
```
## Thinking mode
Ollama now supports thinking mode.
To enable thinking mode in Ollama, first download models that sup... https://github.com/ollama/ollama/releases/tag/v0.9.0
@wrzlbrnft Die Mistral Modelle sind in der Tat recht gut, besonders die Mixtral Varianten. Das Beste an denen ist allerdings, dass man die mit #Ollama problemlos lokal laufen lassen kann, also ohne #Klaut
Convert HTML content to Markdown using LLM and Ollama:
https://www.glukhov.org/post/2025/05/html-to-markdown-using-llm/
#HTML #Markdown #LLM #Ollama
New Ollama Release!
Version: v0.8.0
Release Notes:

## What's Changed
* Ollama will now stream responses with tool calls
* Logs will now include better memory estimate debug information when running models in Ollama's engine.
## New Contributors
* @hellotunamayo made their first contribution in https://github.com/ollama/ollama/pull/10790
**Full Changelog**: ... https://github.com/ollama/ollama/releases/tag/v0.8.0
#Clara - Fully Local #AI Superstack with #Ollama #LLM Chat, Agent Builder & #StableDiffusion
Complete #opensource #AI workspace running 100% offline on your machine
Features #LLM chat with #Ollama models, visual app builder, and #N8N automation integration
Here's a screen cast of the gemma3 1b model running in Ollama in a LXD container on Ubuntu Touch on my Pixel 3a device.
Am Mac verwende ich aktuell #raycast und GPT-4.1 mini um mir eine Bildbeschreibung zu meinen Fotos erstellen zu lassen. Das klappt gar nicht mal so schlecht. Jetzt habe ich mal wieder #ollama angeworfen. Mit Qwen3 oder Gemma 3 geht es schon, aber kommt noch nicht an GPT-4.1 mini ran.
Welches Modell würde sich denn für eine solche Aufgabe anbieten? Danke!
手軽に最新LLM!OllamaとQwen3でローカルAI環境を構築する方法
https://qiita.com/Nakamura-Kaito/items/a135d2c5ecef753ec8be?utm_campaign=popular_items&utm_medium=feed&utm_source=popular_items
New Ollama Model Release!
Model: devstral https://ollama.com/library/devstral
From Text to POJO: Extracting Structured Data with Quarkus, Langchain4j, and Local LLMs
Learn how to turn messy text into clean Java objects using Quarkus, Langchain4j, and Ollama. No APIs, no cloud, just pure local LLM power!
https://myfear.substack.com/p/structured-data-llm-quarkus-langchain4j
#Quarkus #Java #AIML #LangChain4j #Ollama