Mauro > 27-06-2025, 12:18 PM
(27-06-2025, 11:34 AM)Koen G Wrote: You are not allowed to view links. Register or Login to view.So that's more like a polyalphabetic cipher with frequent shifting between alphabets. Sounds like a headache!
nablator > 27-06-2025, 12:57 PM
(27-06-2025, 11:31 AM)quimqu Wrote: You are not allowed to view links. Register or Login to view.a can be ciphered by (T,O,P)
b can be ciphered by (U,P,W)
c can be ciphered by (T,M,Z)
Koen G > 27-06-2025, 01:15 PM
(27-06-2025, 12:18 PM)Mauro Wrote: You are not allowed to view links. Register or Login to view.But I also agree with @Jorge_Stolfi here: n-grams entropies do not necessarily mean much because they are heavily dependent on the transcription. Word entropies are more reliable (they only depend on 'space' being actually a 'space', and on differently written words being actually different words).
quimqu > 27-06-2025, 01:17 PM
(27-06-2025, 12:57 PM)nablator Wrote: You are not allowed to view links. Register or Login to view.(27-06-2025, 11:31 AM)quimqu Wrote: You are not allowed to view links. Register or Login to view.a can be ciphered by (T,O,P)
b can be ciphered by (U,P,W)
c can be ciphered by (T,M,Z)
This is something completely different than the null generator code that only inserted 0, 1, 2. There may be a huge misunderstanding here. Can you post the Python code for the actual mapping, please?
quimqu > 27-06-2025, 01:29 PM
(27-06-2025, 01:15 PM)Koen G Wrote: You are not allowed to view links. Register or Login to view.(27-06-2025, 12:18 PM)Mauro Wrote: You are not allowed to view links. Register or Login to view.But I also agree with @Jorge_Stolfi here: n-grams entropies do not necessarily mean much because they are heavily dependent on the transcription. Word entropies are more reliable (they only depend on 'space' being actually a 'space', and on differently written words being actually different words).
I don't agree with Stolfi's dismissal of letter-based statistics. We don't know if Voynichese is a phonetic representation at all. Whatever it is, is very much dependent on its writing system. So to get a clue about what's going on, we must compare different ways of transliterating Voynichese to different writing systems. Including codes and ciphers, which are essentially also writing systems.
What I do vehemently agree with is that people must be more aware that Voynichese =/= EVA. When running letter-based tests, there must always be awareness of the choices made and their effects.
kckluge > 27-06-2025, 06:22 PM
(27-06-2025, 01:15 PM)Koen G Wrote: You are not allowed to view links. Register or Login to view.(27-06-2025, 12:18 PM)Mauro Wrote: You are not allowed to view links. Register or Login to view.But I also agree with @Jorge_Stolfi here: n-grams entropies do not necessarily mean much because they are heavily dependent on the transcription. Word entropies are more reliable (they only depend on 'space' being actually a 'space', and on differently written words being actually different words).
I don't agree with Stolfi's dismissal of letter-based statistics. We don't know if Voynichese is a phonetic representation at all. Whatever it is, is very much dependent on its writing system. So to get a clue about what's going on, we must compare different ways of transliterating Voynichese to different writing systems. Including codes and ciphers, which are essentially also writing systems.
What I do vehemently agree with is that people must be more aware that Voynichese =/= EVA. When running letter-based tests, there must always be awareness of the choices made and their effects.
cvetkakocj@rogers.com > 27-06-2025, 07:28 PM
obelus > 27-06-2025, 09:43 PM
(27-06-2025, 10:34 AM)quimqu Wrote: You are not allowed to view links. Register or Login to view.I measured n-gram entropy per word
quimqu > 27-06-2025, 10:16 PM
(27-06-2025, 09:43 PM)obelus Wrote: You are not allowed to view links. Register or Login to view.It is tempting to assign "entropy" to a single text sample (which I might call a "micro-message"). But the sequence of characters is known, so all of the probabilities are unity; insofar as it can be said to have an entropy, the value is exactly trivially zero. The meaningful entropy that we discuss here should be attributed to the text generator ("macro-messenger"), as a logarithmic measure of how many different micro-messages it is capable of producing. An approximate value can be calculated from one micro-message, but the accuracy of approximation depends on its length. I expect that there are formulas in the literature that relate n-mer order, sample size, and systematic error. The only way to evaluate high-order conditional entropies of the VMS author is to discover more text that it produced.
Mauro > 27-06-2025, 10:37 PM
(27-06-2025, 01:15 PM)Koen G Wrote: You are not allowed to view links. Register or Login to view.(27-06-2025, 12:18 PM)Mauro Wrote: You are not allowed to view links. Register or Login to view.But I also agree with @Jorge_Stolfi here: n-grams entropies do not necessarily mean much because they are heavily dependent on the transcription. Word entropies are more reliable (they only depend on 'space' being actually a 'space', and on differently written words being actually different words).I don't agree with Stolfi's dismissal of letter-based statistics. We don't know if Voynichese is a phonetic representation at all. Whatever it is, is very much dependent on its writing system. So to get a clue about what's going on, we must compare different ways of transliterating Voynichese to different writing systems. Including codes and ciphers, which are essentially also writing systems.
(27-06-2025, 01:15 PM)Koen G Wrote: You are not allowed to view links. Register or Login to view.What I do vehemently agree with is that people must be more aware that Voynichese =/= EVA. When running letter-based tests, there must always be awareness of the choices made and their effects.I vehemently agree too!!!!!