-
Archives
- April 2024
- March 2024
- December 2023
- November 2023
- October 2023
- July 2023
- April 2023
- March 2023
- February 2023
- October 2022
- February 2022
- January 2022
- December 2021
- September 2021
- August 2021
- July 2021
- March 2021
- January 2020
- May 2018
- May 2017
- April 2017
- March 2016
- February 2016
- December 2015
- September 2015
- March 2015
- January 2015
- October 2014
- June 2014
- March 2014
- February 2014
- December 2013
- November 2013
- August 2013
- July 2013
- June 2013
- May 2013
- February 2013
- October 2012
- July 2012
- May 2012
- April 2012
- March 2012
- February 2012
- January 2012
- November 2011
- October 2011
- June 2011
- May 2011
- April 2011
- December 2010
- November 2010
-
Meta
Category Archives: Large Language Models
Securing large language models with a reverse proxy
In a previous post, I explained how to host a private ChatGPT using Docker and Traefik. I didn’t spend a lot of time on the security aspect of the project. I see many people asking how to expose their large … Continue reading
Self-hosted coding assistant with llamafile, continue.dev and docker
There was a recent dramatic improvement on the speed of LLM’s on CPU thanks to llamafile‘s author. She goes on extensively about it on her blog but the short version is: expect 7-billion parameters to be usable on consumer-grade CPU … Continue reading
Europe GPU prices update – March 28 2024
With all the buzz about AI these days, let’s have a look at the GPU prices in Europe and check which one gives the best “bang for the buck” as YouTubers like to say. YouTube is filled with people telling … Continue reading
Ollama, open-webui, mitmproxy in a docker compose stack, behind traefik
Reading Ollama discord channel, I notice many people want to self-host their chatGPT with Docker and don’t know how to do it. Here’s how to host the whole stack with docker compose.
Ollama system prompt
I have recently started to use Ollama and I was unimpressed by some models as they did not follow instructions, especially in their output format. I knew about model system prompt but I thought it was fixed in the model. … Continue reading