LLMs are understood to generate non-deterministic outputs. My question is wether there are LLMs out there that are capable to producing deterministic outputs for any given input given fixed parameters (like e.g temperature).
I heard that llama.cpp - if run on a CPU instead of a GPU appears to generate deterministic outputs.