geoffreycaveney > 20-09-2020, 02:57 AM
(19-09-2020, 09:49 PM)RenegadeHealer Wrote: You are not allowed to view links. Register or Login to view.(19-09-2020, 02:42 PM)geoffreycaveney Wrote: You are not allowed to view links. Register or Login to view.I want VCI transcription to be a tool that all Voynich researchers can use. The idea of [f]/[p] as a variant of [d] rather than of [k]/[t] had come to feel like a "pet" theory of mine last year, so I don't want to impose such a minority opinion of mine on a transcription system designed for general use.
Thanks for clarifying. I guess I didn't realize that you custom built the idea [f]/[p]=[d] in service to your Judeo-Greek theory. I thought it was an idea with some merit, independent of any idea of what the language might be, or even whether Voynichese is language at all.
Quote:Last winter I was actually inspired by your idea to design a preliminary test of the hypothesis that [p] is a top-line-only equivalent of some other Voynichese glyph or string of glyphs. I describe the test here:
You are not allowed to view links. Register or Login to view.
I'm going to run it when I get over the hump of learning to work with and statistically analyze VMs transcriptions, and post the results when I do.
Quote:On the subject of a tool for general use, what do you foresee other VMs researchers using your VCI system to do? I can tell you what I plan on using it for: cribbing. One of these days I'm going to pick out some strategically placed lines of Voynichese that have been of interest as potential crib sentences due to their context with the imagery, and try using your tool to convert it to the graphemes you suggest. I'll then see if I feel the slightest activity in my hippocampus — the smallest sensation of "this feels vaguely familiar", and trying to place it. If this is an actual promising lead, and not just wishful thinking with too many degrees of freedom, it should soon lead to more Voynichese text becoming comprehensible. If nothing I transformed with your (or anyone else's similar) tool made me say "aha", or none of the "aha" moments swiftly led to bigger "aha"s, then I'd have no further use for your tool. Is that more or less how you were envisioning others using your tool?
Quote:Quote:seems clear that the general working assumption of most Voynich researchers has been that [f]/[p] are most likely variants of [k]/[t], so VCI is faithful to that consensus.
What's your source for this? I'm surprised to hear you say this, because it's my understanding that this is far from a consensus. It's not an uncommon opinion, but it's not supported by the occurrence of all gallows (and [d]) together on some lines. A good example is You are not allowed to view links. Register or Login to view. line 12.
Quote:Quote:At this stage I don't want to force too many such n-gram=single unit equivalences into the system. The system already has plenty of them. As it currently stands VCI can read [dch] as <ki>, and I don't want to get rid of any more vowels than I have to. Of course if the text ends up being Arabic or Maltese, then maybe we don't need all those vowels. But if it ends up being Czech or Irish or Basque, then we should expect to see a normal amount of vowels. Also, it is easy enough to treat <ki> as <kj> at a later stage, if we want to go in that direction.
Fair enough. I'm just curious, and looking to get more insight into how you built the model. I'm sorry if I missed this, but based on Koen and Marco's work, did you define numerical parameters for where you made the cutoff of which ngrams to treat in your model as single units, versus which ones to continue treating as strings of n units each? If this were my project I probably would, just to make my tool more user-friendly and transparent.
Quote:Quote:I think EVA [y], which is also VCI <y>, is a very important and complicated glyph in the ms text. I would not want to rush to equate it with EVA [a] and potentially lose essential distinct information that [y] actually contains and represents. If in the end [a] and [y] do prove to be equivalent, it will still be possible to detect that at a later stage in due time: "Linguy Latiny per se Illustraty, Pars I: Familiy Romany" is not a difficult cipher step to figure out. But if we equate them now, and [y] proves to be distinct, it will be more difficult to recover that distinction if we are all using a system that treats them and presents them as identical.
While we're at it, since I asked you for a source about the ideas [f]=[k] and [p]=[t], I should mention that my source for the idea [a]=[y] is Emma May Smith's blog. Like the equivalences your model allows, this one is controversial, and not at all consensus. But Emma gives some tantalizing clues that it might be true. I only mention this because I'm going to be keeping this possible equivalency in mind while trying out your tool, and I know you're on the lookout for possible glyph equivalencies to make your model better.
Quote:Quote:Yes, I think I said in the post accompanying the VCI tables that [al] was the most difficult decision. The system is more internally consistent if EVA [al] = VCI <as>. But treating the bigrams [ol], [or], [al], [ar] as each representing a single unit was part of Koen's method in generating the 3.01 conditional entropy value for the ms text, and I aimed to have the VCI system respect that method as much as possible. I couldn't find any consistent way to force the treatment of [ar] as a single unit, so I let that go as <al>. (Please note that I was not even remotely thinking of Arabic when I made this decision--if anything, I was thinking of Slavic past tense verb endings!) But out of respect for Koen's method and raising conditional entropy, I forced EVA [al] = VCI <a> in the spirit of his verbose cipher analysis.
Thanks for taking the time to indulge my "couple" of questions, Geoffrey. I think the most challenging part of attempting to reverse engineer an unknown system for encoding human phonemes, is determining the right amount of detail and complexity to add before trying it. Add too little, and the system has too much ambiguity and too many degrees of freedom. Add too much, and chances are a lot of the details are wrong and will lead you astray. Also, spend too much time and energy building a model, and it's easy to get too emotionally invested in your masterpiece to accept it not working.
Aga Tentakulus > 20-09-2020, 07:04 AM
Aga Tentakulus > 20-09-2020, 07:08 AM
Koen G > 20-09-2020, 10:18 AM
Emma May Smith > 20-09-2020, 12:06 PM
ReneZ > 20-09-2020, 01:02 PM
nickpelling > 20-09-2020, 01:12 PM
(19-09-2020, 10:17 PM)Emma May Smith Wrote: You are not allowed to view links. Register or Login to view.By "presence of common groups" do you mean the bigrams themselves or that such bigrams are meaningful units?
Koen G > 20-09-2020, 01:46 PM
ReneZ > 20-09-2020, 02:00 PM
(20-09-2020, 01:46 PM)Koen G Wrote: You are not allowed to view links. Register or Login to view.It is definitely certain that Voynichese was not meant to be transformed in this exact way
geoffreycaveney > 20-09-2020, 02:15 PM
(20-09-2020, 01:02 PM)ReneZ Wrote: You are not allowed to view links. Register or Login to view.Back on the topic of this thread, the work by Koen and Marco was of course intended to see how one can bring the entropy of the MS text to a level compatible with known plain texts by combining characters, i.e. trying to identify components of a verbose cipher. In the course of that, they had many dozens of different 'alphabets', as the plots at Koen's blog shows.
As far as I recall, this iterative process did not include the option to 'undo' earlier combinations as a next step. If that would be allowed, the number of different 'alphabets' grows even further. Another thing they did not do was to consider certain characters as equivalent. Doing that really explodes the number of possibilities.
If one would allow all that, indeed a genetic algorithm seems like the most suitable approach.
Now coming to the VCI 'alphabet', this is not a transcription / transliteration, but an interpretation, and it is based on just one of the above-mentioned myriad of possible 'alphabets', all of which are equally (im-)probable.
It is based completely on speculation.