|Title||Long Short Term Memory (2017)|
|Media||Hardcover Book (92 pp, 8 x 10 in, Edition of 5)|
Long Short Term Memory (2017) comprises texts and images produced by artificial neural nets imbued with memory, deploying architectures for forgetting within the realm of machine learning.
A recurrent neural net (RNN) trained on a corpus of contemporary rationalist philosophy and computational theory makes its own synthetic claims. Devoid of grammatical axioms, language is but a stream of characters to the RNN. Ingesting the work of Luciana Parisi, François Laruelle, Quentin Meillasoux, and others, the net offers its own novel speculations.
Alongside, the latent space of an LSTM network is hijacked with abstract inputs, revealing formal aspects of its architecture in a series of black & white exposures. These allude to smooth activation functions arranged in such a way as to decay information during the process of learning.
|License||© All Rights Reserved|