Replies: 9 comments 22 replies
-
@akolonin I tried to define transition freedom and peak freedom in a formal fashion, in the attached PDF. Is this correct? Hmm seem I cannot attach, I'm getting github error messages. The PDF is here: peak freedom PDF please comment. |
Beta Was this translation helpful? Give feedback.
-
Hi Linas, I agree that the paper by Wrenn et.al. is pretty vague, speaking softly ;-) Notice, that the "peak" works good only for Chinese, the "variance" is better for English and Russian. |
Beta Was this translation helpful? Give feedback.
-
@nickm197 @linas - we have extended the INLP workshop deadline https://aigents.github.io/inlp/2022/ |
Beta Was this translation helpful? Give feedback.
-
Indeed, one can look at it this way
Who is not these days?
This setting is the exact one where the classical Popperian discourse is valid (and it is valid until the knowledge field becomes so entangled that finally you can only falsify - in popperian terms - the whole mentis): You work with hypotheses and keep them falsifiable. In other words, there is no actual truth, but a succession of models that increasingly better describe reality. A model is an explanation of reality that accepts a certain viewpoint of the reality and, thus, compresses information about reality with losses. Occam's razor is a requirement to improve compression of information in a model.
I am in no way an expert in genetic algorithms, but always considered them to be a descent in a rather fancy discrete space. Thus, the result depends on the way you construct this space more then on anything else. In math, one would need to prove that the optimal solution belongs to that space. |
Beta Was this translation helpful? Give feedback.
-
Hi @linas can you email me your slides for your talk on August 19? |
Beta Was this translation helpful? Give feedback.
-
@nickm197 we have got potentially relevant reference to our work https://www.researchgate.net/publication/361505964_Estimating_Sentence-like_Structure_in_Synthetic_Languages_Using_Information_Topology/references |
Beta Was this translation helpful? Give feedback.
-
Hi Nick, I want to continue the conversation but have been distracted by other matters. Are you in (or would you be able to get to) Europe next month? |
Beta Was this translation helpful? Give feedback.
-
@linas @nickm197 we will meet on Aug 19, check your email and workshop agenda - need your slides by then ;-) https://aigents.github.io/inlp/2022/ |
Beta Was this translation helpful? Give feedback.
-
@linas - you are about to present on Aug 19 at https://aigents.github.io/inlp/2022/ and on Aug 22 at http://agi-conf.org/2022/schedule/ |
Beta Was this translation helpful? Give feedback.
-
The discussion in #1310 wanders off into tokenization, and a distinct thread for that seems best. Specifically this comment from @akolonin
I just read the following:
An Unsupervised Machine Learning Approach to Segmentation of Clinician-Entered Free Text
AMIA Annu Symp Proc. 2007; 2007: 811–815.
PMC2655800
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2655800/
I found it frustratingly opaque and imprecise. Work items and clarifications suggest themselves:
informal definition is imprecise, prone to misunderstanding, and obscures
relationships to other similar mathematical formulas.
difference between the prior and the subsequent transition freedom.
What if, instead, one worked with log2 of transition freedoms, so
that a peak entropic freedom was a ratio instead of a difference?
how are they related to more traditional entropic definitions?
thousands of transitions; what's the freedom of each? What's the peak
freedom of each?
the next pair of characters, or the next triple of characters?
@akolonin by "distribution" I mean this: consider taking the bar charts on page 5-6 of your paper (the https://arxiv.org/abs/2205.11443 one) and projecting them onto the right axis. This gives a distribution that shows how often any given height occurs. The questions: what's that distribution? Is it gaussian? Something else? What's the mean and deviation? ... Personally, I cannot get comfortable with any of this stuff until I see the distributions, and also the formal definitions, and the variants that use log2 of transition freedoms.
The Wrenn etal 2007 paper and the Kolonin 2022 paper seem to work with "transition freedoms" by exhaustively counting all possible transitions between substrings and single characters. What if instead, we worked with transitions between substrings and substrings? There are explosively many of these, and so perhaps a random subsampling is more computationally feasible? This is why I mentioned the "amy" language in #1310 (comment) : it provides a random sampling of transitions.
Beta Was this translation helpful? Give feedback.
All reactions