It looks like you're using an Ad Blocker.
Please white-list or disable AboveTopSecret.com in your ad-blocking tool.
Thank you.
Some features of ATS will be disabled while you continue to use an ad-blocker.
I am looking to chat with other ATS'ers who have a background with AI, deep learning, big data, etc. Personally I am interested in hybrid deep learning, and video applications combining static visual, short term patterns, and long term temporal clues however this thread is open to all AI topics ranging from common to esoteric.
During reinforcement learning, an agent A attempts to improve its parameters from conversations with another agent B. While the other agent B could be a human, in our experiments we used our fixed supervised model that was trained to imitate humans. The second model is fixed as we found that updating the parameters of both agents led to divergence from human language. In effect, agent A learns to improve by simulating conversations with the help of a surrogate forward model.
6.3 Intrinsic Evaluation For development, we use measured the perplexity of user generated utterances, conditioned on the input and previous dialogue. Results are shown in Table 3, and show that the simple LIKELIHOOD model produces the most human-like responses, and the alternative training and decoding strategies cause a divergence from human language. Note however, that this divergence may not necessarily correspond to lower quality language—it may also indicate different strategic decisions about what to say. Results in §6.4 show all models could converse with humans.
originally posted by: Phantom423
Any opinions on this research?
originally posted by: Aazadan
originally posted by: Phantom423
Any opinions on this research?
It basically shows how languages evolve over time. Humans do the same thing but on a slower scale. English from 1000 years ago would be completely incomprehensible to you today.
All reinforcement learning models do this, it's one of the downfalls of the algorithms.