Skip to content
No results
  • Models
    • Analytical Reasoning Llama
    • HermesNova
    • Brainstorm 3.1
    • Brainstorm (Llama3)
    • Brainstorm Plus (Llama3)
    • Brainstorm-v2.1 (Llama3)
    • Brainstorm (Falcon2)
  • Software
    • LLmaaS
    • LLMForwarder
    • LLM Predictive Router
    • No More Typo
    • aiHub Mobile iOS app
    • aiHub
  • Misc
    • GPU GGUF inference comparison
  • Blog
  • Contact
DevQuasar
  • Models
    • Analytical Reasoning Llama
    • HermesNova
    • Brainstorm 3.1
    • Brainstorm (Llama3)
    • Brainstorm Plus (Llama3)
    • Brainstorm-v2.1 (Llama3)
    • Brainstorm (Falcon2)
  • Software
    • LLmaaS
    • LLMForwarder
    • LLM Predictive Router
    • No More Typo
    • aiHub Mobile iOS app
    • aiHub
  • Misc
    • GPU GGUF inference comparison
  • Blog
  • Contact
DevQuasar
  • AI, dataset, diffusion, prompt, sythetic, vision

Synthetic-Cyclic-Perception

I wanted to experiment with combining image generation and vision models to see how images evolve when each description is used as the next prompt for the generator. My goal was to observe whether this feedback loop causes images to…

  • Csaba Kecskemeti
  • October 27, 2024
  • Edge

Satisfying gif – Stablediffusion text to image steps

Sample code to generate the gif

  • Csaba Kecskemeti
  • October 19, 2024
  • AI, Chat with document

An interesting Chome plugin: elmo.chat

Elmo chat is Lepton AI’s little fun chrome extension that lines in your chrome browser and offering summarization question answering and even TTS (Text To Speech) do chat with the webpage you’re browsing. Slowly but surely changing the way we’re…

  • Csaba Kecskemeti
  • October 17, 2024
  • Edge, Inference, locaal, Uncategorized

Speculative decoding tradeoffs GPU with Large models (Llama 70B & 8B)

I’ve rerun the Speculative Decoding experiment with some larger models where I’ve paired a Llama 3 70B primary model with the 8B draft model, to see if a larger model Llama 3 70B can benefit more from a draft model…

  • Csaba Kecskemeti
  • October 12, 2024
  • Edge, Inference, locaal, Uncategorized

Speculative decoding tradeoffs GPU

And there I’ve crunched the numbers on GPU. The configuration is a 3GPU system 1x RTX4080 + 2x RTX3090. Baseline has been set by the following llama.cpp generation command (same generation config and prompts has been used in the Speculative…

  • Csaba Kecskemeti
  • September 24, 2024
  • AI, Edge, Inference, locaal, Uncategorized

Speculative decoding on CPU tradeoffs 2

Here’s a followup on the Speculative decoding on CPU tradeoffs now performing the same test on a Threadripper 3970x, and finally I see some benefit. Acceptance rate Generation speed Summary The results seems very similar to the Xeon E5 test.…

  • Csaba Kecskemeti
  • September 23, 2024
  • AI, Edge, Inference, locaal

Speculative decoding on CPU tradeoffs

I’ve played around with the llama.cpp speculative decoding on CPU (Mac Pro 2013 – Xeon E5 12core 2.7GHz) and wanted to share my experience. First of all I’ve struggled to find models where the vocab size difference is less than…

  • Csaba Kecskemeti
  • September 21, 2024
  • prompt, Software

NO Code Galaga game in 10 prompts using Llama 3.1

After initial prompt it’s already working, quite impressive After 10th prompt it look like a game

  • Csaba Kecskemeti
  • September 10, 2024
  • AI, Edge, Edge, Inference, locaal, Uncategorized

The cheapest way to host a local LLM – with reasonable speed

That’s probably the trash can Mac Pro from 2013. In December 2013, Apple released a new cylindrical Mac Pro (colloquially called the “trash can Mac Pro). Apple said it offered twice the overall performance of the first generation while taking…

  • Csaba Kecskemeti
  • September 6, 2024
  • Hardware, locaal, Uncategorized

ML Workstation upgrade – 2024 Summer

I’ve decided to upgrade my ML workstation, I was especially interested to improve the CPU core count. Unfortunately the original Asus ROG Zenith Extreme only supports Threadripper Gen2 processors and as 2990WX is more expensive than a 3970X which is…

  • Csaba Kecskemeti
  • August 11, 2024
Prev
1 2 3 4
Next

Copyright © 2025 devquasar