About removal of classical evaluation. #4678
Replies: 4 comments 11 replies
-
I am curious about the implication about the whole information loop in the NNue Master network training, as it might depart from the available blog from SF12 (to my knowledge, i would like to be informed otherwise with link to follow) where it is made clear that the classical evaluation version of SF at that version, was the "orable" trainer for the NN to approximate on some big position set. I think I have not missed one blog version page and did not see that SF12 blog explanation had been changed, but with SF16, some new language minimally referring to leela's data was used. My understanding for now, or hypotheses of understanding, are that minimally, it is about using the training games positions. There are other ways to use leela's data that might be a structure change in my opinion about the whole information flow in the design of SF evolution (sorry for the word choice is troubling). But it might requires a minimal bit of information other than how fast the implementations have become.. It is a matter of user interpret-ability of the tool. And it might be very easy to do for those who know. And it might not be the prison that good documentation other than source code might have haunted the developer imagination in open source collaborative projects (if that is the resistance). I know chess likes to trade in secrets and expertise auras, but we are talking about programmable things. Someone ought to have the perspective to answer my question. If using more than the training games databases ground into position database, and keeping the whole game outcome information as part of the SF master network training. . Then what kind of machine learning procedure is used. How are the games outcome integrated. Could the isolated repository about WDL conversions from one of the SF developers have something to do with it. How? but the simplest question is, does SF12 blog explanation of the outer flow of information on the network training coming from classical searches still stand.. I am writing here because I would like the op to please provide some links to where such decision might have been discussed. As it does not seem compatible with SF12 blog. and the absence of blog announcements that would modify that. I am sorry if missed a blog paragraph though. I think source code is not use documentation. So, it might be a bit from exasperation about that possible tendency, that I might sound a bit tense about this issue.. but really i am deeply interested by how training is done, more than how fast the executable performs.. I think documentation might be a bit more transparent about that. it does not require developer inside knowledge to be able to understand such level of reading. It has nothing to do with quantization (well not directly) or feature reduction at input (well not directly). It can be shared perhaps at the same minimal disclosure level as SF12 blog. I would congratulate such effort in documenting something so important for interpretation by serious users and all lichess users for example. |
Beta Was this translation helpful? Give feedback.
-
Although I'm new incomer here but from a chess programmer's point of view, we can't remove the classical eval until at least 90% of the chess positions have been trained. Maybe it would be better if this removal was done in Stockfish 20. |
Beta Was this translation helpful? Give feedback.
-
First, I have to see if this problem will be solved by structural changes in network training or not. Recently, I opened a discussion asking what is the difference between a network that is trained with @cj5716 Can you help me in this context? |
Beta Was this translation helpful? Give feedback.
-
I think this is a big step foward to make it in a rush. I´m not saying it is wrong but I would say it could wait a little longer, when the classical evaluation would not bring any real elo to the engine (proved). It´s maintanance were not really "harming" the engine progress. No "nostalgic" feelings about it. It seems this code would not be removed if treated as a normal "simplifying test" as it should. I think it pragmatically failed as a "simplyfication test" by SF´s stardard rules but made by "The Director´s will". (Do you understand that about 3 ELO in this level maybe represents about +15 or even more in other times? This is due to "shrinking gains" in high level of play. Did you consider this?)
Anyway...as I said...a big step foward, maybe it is reasonable, hope it will bring us a stronger engine in the long term. I miss the time the focus of SF Project was +ELO. Let´s hope the best! Good luck and job.
Beta Was this translation helpful? Give feedback.
All reactions