@drewdaniels
yes very different. PPM operates of surface characteristics of text output. Seems to be getting more difficult as LLMs get more sophisticated.
My work came about from examining other LLM models (not the ones used for generation). I guess the research best described as "how other LLMs react” to text. With the hypothesis being that weird LLM behaviour has been shown in LLMs trained recursively on LLM output (i.e. models trained on model output rather than authentic human text, shown by Shumailov et. al, 2024 - https://www.nature.com/articles/s41586-024-07566-y). IT was motivated by me trying to automate recipe parsing from blogs (to extract ingredients for my shopping cart), but getting weird hallucinations if the copy-paste to my LLM tool came from another LLM!
My experience made me think what is it that differs in the way a transformers layers (residual state, Attention, MLP) are activated when processing real human text vs. LLM text.. the method is about selecting the families and internal characteristics that allow this).
Edit: correcting typos and clarity