I read your post (actually a few of them) and I have some questions:
- I was surprised to read about n-gram model because at least in NLP was outperformed by modern DL approaches. But then I read you considered RNN (and maybe attention model?) so it sounds really interesting.
- ERNIE 2.0 supports continual learning (AFAIK this was only applied to NLP) so develop something which learns incrementally would be an option for your PhD?
One problem, for me, of code prediction is gathering enough context information to make informed predictions (I ignore if there is something like GLUE benchmarks for code). This means to propose contextualized completions based on multiple factors: for example the class where I am positioned in the Browser, the recent used/written methods, etc. Of course the possibilities are infinite! Just thinking out loud.
El lun., 2 dic. 2019 a las 11:25, Myroslava Romaniuk via Pharo-users (<[hidden email]>) escribió: