Jorge_Stolfi > 25-08-2025, 08:30 PM
dashstofsk > 25-08-2025, 08:45 PM
(25-08-2025, 01:45 PM)magnesium Wrote: You are not allowed to view links. Register or Login to view.interesting to sweep across a wider parameter space for this generator: randomly iterate the code’s various threshold parameters, sweep across a wide range of initializing lines
asteckley > 25-08-2025, 09:09 PM
(25-08-2025, 08:45 PM)dashstofsk Wrote: You are not allowed to view links. Register or Login to view.It seems to me unlikely that any tinkering with the algorithm will lead to any leap in understanding of the manuscript. If you read Torsten Timm's papers you will read that his 'simple process for random text generation' just reproduces 'the key statistical properties' of the language of the manuscript. There is no claim that his method can truly reproduce the text of the manuscript.
ReneZ > 26-08-2025, 12:16 AM
Torsten > 27-08-2025, 03:33 PM
(24-08-2025, 08:46 AM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.(23-08-2025, 08:23 PM)Torsten Wrote: You are not allowed to view links. Register or Login to view.There was no need to invent an artificial “gibberish generation” mechanism. As D’Imperio already observed [...]
She was not stating a fact. She was proposing her version of the "hoax" theory. Which, in general terms, apparently is the same as yours. Which has the same problems as yours.
Torsten > 27-08-2025, 08:02 PM
(24-08-2025, 08:46 AM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.Torsten Wrote:Consequently, a scribe attempting to generate language-like gibberish would, sooner or later, abandon the laborious task of perpetual invention in favor of the far easier strategy of reduplicating and adapting previously written material — and would ultimately adhere to this approach consistently.
Note my emphasis. The problem is that the "adapting" is far from a simple step. Voynichese words have a very restricted structure, so the "adapting" must be random but such that it preserves that structure. At this point the gibberish generation method is not much easier than generating each word from scratch (as Rugg had proposed), and is totally not "natural".
(24-08-2025, 08:46 AM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.In fact (if I read you correctly), your justification for your proposed method is that it creates the repetitiousness that you claim to see in the VMS; which is a clue that the text is gibberish. Wouldn't the Author have worried about this last fact?
Paraphrasing your argument: "The VMS text has statistical properties X, Y, and Z, where Z is 'repetitiousnss'. Here is an algorithm that generates gibberish with properties X, Y and Z. Therefore the VMS must be gibberish."
(24-08-2025, 08:46 AM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.How could a "parameterless" Mutate function produce these asymmetric word frequencies?
Quote:The respective frequency counts confirm the general principle: high-frequency tokens also tend to have high numbers of similar words. This is illustrated in greater detail in Figure 3: "isolated" words (i.e. unconnected nodes in the graph) usually appear just once in the entire VMS while the most frequent token <daiin> (836 occurrences) has 36 counterparts with edit distance 1. Note that most of these "isolated" words can be seen as concatenations of more frequent words (e.g. <polcheolkain>=<pol>+<cheol>+<kain>). This characteristic dependence of token frequency from word similarity is just another manifestation of the long-range correlations that have been uncovered and discussed by several researchers throughout the last decade.