Jorge_Stolfi > Yesterday, 11:59 AM
(Yesterday, 07:51 AM)magnesium Wrote: You are not allowed to view links. Register or Login to view.Why go through all the trouble of making a complex cipher if you could sell the book just as easily with it saying nothing at all?
dexdex > Yesterday, 07:22 PM
Torsten > Yesterday, 08:23 PM
(Yesterday, 11:59 AM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.First, it would be a sure way to make Voynichese look like a real language. It seems unlikely that a forger of the time could devise a gibberish generation method with that property. The methods of Rugg and Thorsten&Timm could do that with proper parameters, but no one has shown how the VMS Author could have figured out the proper parameters.
Quote:We deliberately did not fine-tune the algorithm to pick an 'optimal' sample for this presentation. Such a strategy is by itself questionable. Nevertheless, an exhaustive scan of the parameter space (involving thousands of automatically analyzed text samples) verified the overall stability of the proposed algorithm. About 10-20% of the parameter space even yields excellent numerical conformity (≤ 10% relative error) with all considered key features of the real VMS text (entropy values, random walk exponents, token length distribution, etc.).
dexdex > Yesterday, 09:03 PM
(Yesterday, 08:23 PM)Torsten Wrote: You are not allowed to view links. Register or Login to view.One of the key advantages of the self-citation method is that statistical regularities—such as the Zipfian distribution—emerge organically from the process itself, without requiring deliberate adjustment. This is because the method is grounded directly in observations of the Voynich text. I begin by identifying fundamental, corpus-wide patterns—for example, the clustering of similar words across pages. These clusters suggest a mechanism of repetition and gradual variation. The central argument, therefore, is that an iterative process of copying and modification is sufficient to account for the statistical features. The observed word type frequencies can be interpreted as the outcome of a self-reinforcing resonance effect produced by continual copying. Frequently occurring words are more likely to be copied, thereby generating numerous variants. The accumulation of such variants, in turn, increases the probability that the original form will reappear during subsequent cycles of copying. This way the hypothesis accounts for the clustering of similar words within the Voynich Manuscript.Furthermore, in regards to "no one has shown how the VMS Author could have figured out the proper parameters." - the author(s) didn't need to 'figure out' the parameters. Had they accidentally landed somewhere else in parameter space, we would be very similarly stumped. All that's required for the hypothetical process to be plausible is that the fraction of the parameter space that explains the Voynich is significant enough that they could have landed there. Which it sounds like this simple process is.
Therefore there is no need to determine or fine-tune external parameters. As noted in Timm & Schinner (2019):
Quote:We deliberately did not fine-tune the algorithm to pick an 'optimal' sample for this presentation. Such a strategy is by itself questionable. Nevertheless, an exhaustive scan of the parameter space (involving thousands of automatically analyzed text samples) verified the overall stability of the proposed algorithm. About 10-20% of the parameter space even yields excellent numerical conformity (≤ 10% relative error) with all considered key features of the real VMS text (entropy values, random walk exponents, token length distribution, etc.).
Jorge_Stolfi > 1 hour ago
(Yesterday, 08:23 PM)Torsten Wrote: You are not allowed to view links. Register or Login to view.There was no need to invent an artificial “gibberish generation” mechanism. As D’Imperio already observed [...]
Torsten Wrote:It is simply more efficient for a scribe to copy and modify existing words than to continually invent entirely new ones.
Torsten Wrote:Consequently, a scribe attempting to generate language-like gibberish would, sooner or later, abandon the laborious task of perpetual invention in favor of the far easier strategy of reduplicating and adapting previously written material — and would ultimately adhere to this approach consistently.
Torsten Wrote:I begin by identifying fundamental, corpus-wide patterns — for example, the clustering of similar words across pages. These clusters suggest a mechanism of repetition and gradual variation. The central argument, therefore, is that an iterative process of copying and modification is sufficient to account for the statistical features. [...]
def TnT(Seedtext,Mutate,Prob_Restart,Prob_Mutate):
# Generates a pseudo-VMS text as a list of strings.
#
T = SeedText(1000) # Create a 1000-word seed text.
k = None; # Source text index.
for i in range(35000):
if i == 0 or random() < Prob_Restart:
k = randint(0, len(T)-1)
word = T[k]; k += 1
if random() < Prob_Mutate:
word = Mutate(word)
T.append(word)
return(T)
Torsten Wrote:Therefore there is no need to determine or fine-tune external parameters
(Yesterday, 09:03 PM)dexdex Wrote: You are not allowed to view links. Register or Login to view.the author(s) didn't need to 'figure out' the parameters. Had they accidentally landed somewhere else in parameter space, we would be very similarly stumped. All that's required for the hypothetical process to be plausible is that the fraction of the parameter space that explains the Voynich is significant enough that they could have landed there. Which it sounds like this simple process is.
dexdex > 1 hour ago