ReneZ > 11-09-2025, 08:28 AM
oshfdk > 11-09-2025, 08:42 AM
(11-09-2025, 08:19 AM)quimqu Wrote: You are not allowed to view links. Register or Login to view.The results I ppsted to René comment about entropy are the results of decoding the full De docta ignorantia text.
Yes, I have the code. Need to clean it a bit, and publish. I would like to publish a paper about this together wirh the code, but firstly I wanted some feedback from the ninja comunity if it is really interesting or not.
nablator > 11-09-2025, 08:50 AM
(11-09-2025, 08:10 AM)quimqu Wrote: You are not allowed to view links. Register or Login to view.Yes, the tricky thing is to adapt it to voynich stile and try to keep the residuals as low as possible.
quimqu > 11-09-2025, 04:30 PM
(11-09-2025, 08:28 AM)ReneZ Wrote: You are not allowed to view links. Register or Login to view.To get a really good match, one should be able to approach the entire distribution.
Note, that this would be a major achievement, and a sign that one is potentially onto something.
Jorge_Stolfi > 11-09-2025, 05:52 PM
(11-09-2025, 04:30 PM)quimqu Wrote: You are not allowed to view links. Register or Login to view.Ready to try it! But maybe first I should get a text that has a similar distribution of word lengrhs.
quimqu > 11-09-2025, 08:27 PM
(11-09-2025, 08:42 AM)oshfdk Wrote: You are not allowed to view links. Register or Login to view.Nothing wrong with that, but essentially this is the same old idea that the actual alphabet of Voynichese is not ~15 common characters, but ~30-40 common characters with two or more distinct l's, two or more distinct r's, etc.
oshfdk > 11-09-2025, 09:16 PM
(11-09-2025, 08:27 PM)quimqu Wrote: You are not allowed to view links. Register or Login to view.These slight gliph differences could be perfectly the residuals marks, so, no need of a suplementary booklet with the residuals.
quimqu > Yesterday, 10:45 AM
(11-09-2025, 08:28 AM)ReneZ Wrote: You are not allowed to view links. Register or Login to view.On a more detailed note, keep in mind that the second order entropy (either the unconditional or the conditioal one) is a single value representing a 2-dimensional matrix of probability values.
There are infinitely many different such matrices (probability distributions) that all lead to the same entropy.
To get a really good match, one should be able to approach the entire distribution.
Note, that this would be a major achievement, and a sign that one is potentially onto something.
(Of course, a Naibbe-type cipher, that recomposes the text from known segments of it, may be able to mimick it quite well due to its very nature.)
Text | JSD |
---|---|
Vietnamese with special chars | 0.504 |
Vietnamese without special chars | 0.362 |
Romeo and Juliet | 0.299 |
De docta Ignorantia | 0.331 |
In Psalmum David CXVIII Expositio | 0.316 |
Lazarillo de Tormes | 0.340 |
Triant lo Blanch | 0.317 |
Jorge_Stolfi > Yesterday, 12:56 PM
(Yesterday, 10:45 AM)quimqu Wrote: You are not allowed to view links. Register or Login to view.Vietnamese without special chars
Quote:I’ve been comparing full bigram distributions directly. For each text I build the 2D bigram probability matrix (word-reset, normalized), flatten it, and compute the Jensen–Shannon divergence (JSD) against the Voynich matrix.
quimqu > Yesterday, 01:14 PM
(Yesterday, 12:56 PM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.Second, IIUC, the JSD assumes that the two distributions P,Q are defined on the same set of elements, and compares the probabilities assigned by P and Q on each element. Thus if P is the letter pair distribution of a text, and Q is the letter pair distribution of the same text encoded with a Caesar cipher, the JST should come out huge. Do you take this into account, and try to vary the mapping between letters of the two languages?
(Yesterday, 12:56 PM)Jorge_Stolfi Wrote: You are not allowed to view links. Register or Login to view.In case you haven't noticed, those "special chars" are modifiers for the letters. It makes no sense to exclude them.