50 years after the perceptron, 25 years after PDP: Neural computation in language sciences
This Research Topic aims to showcase the state of the art in language research while celebrating the 25th anniversary of the tremendously influential work of the PDP group, and the 50th anniversary of the perceptron. Although PDP models are often the gold standard to which new models are compared, the scope of this Research Topic is not constrained to connectionist models. Instead, we aimed to create a landmark forum in which experts in the field define the state of the art and future directions of the psychological processes underlying language learning and use, broadly defined. We thus called for papers involving computational modeling and original research as well as technical, philosophical, or historical discussions pertaining to models of cognition. We especially encouraged submissions aimed at contrasting different computational frameworks, and their relationship to imaging and behavioral data.