Mike Stay, 9/27/2019, “Transformer Model”... Links to: https://openreview.net/pdf?id=S1eZYeHFDS I still feel the same way: despite an interest in ODEs and function theory, the author’s failure to contextualize in the wider arena limits my interest in the paper. Non-specialists are further excluded by lots of machine learning jargon. Perhaps that is the point. —Brad
On Apr 28, 2020, at 5:51 PM, Marc LeBrun <mlb@well.com> wrote:
Don't recall this getting discussed here: https://arxiv.org/pdf/1912.01412.pdf
Interestingly treats symbolic integration as a kind of linguistic translation. Leverages the cost asymmetry "trap door" versus differentiation when training.
Thoughts?
(Might something similar deconvolve a long digit string into two prime factors? Multiplying up a training set of a couple of hundred million products is easy!)
_______________________________________________ math-fun mailing list math-fun@mailman.xmission.com https://mailman.xmission.com/cgi-bin/mailman/listinfo/math-fun