skip to content
🐈‍⬛ Andrew Zhang

Thought on TTT-E2E

Last Updated:

Just read the TTT-E2E paper and had some informal thoughts. It seems like the critical weakness is hardware efficiency. Due to TTT’s nonlinear recurrence, they must approximate with minibatch TTT. Furthermore, they sequentially backprop through the TTT layers, which is inefficient.

These two weaknesses led me to wonder about a possible research direction. Can we approximate the last hidden state with a simple linear draft model (similar to Eagle 1’s architecture and regression objective) rather than using cross entropy loss across the entire vocab. Then the recurrence would be linear (and possibly parallelizable with chunkwise parallel form) and the backprop would be minimal. Furthermore, to parallelize across layers, hidden states could be extracted from multiple layers, similar to Eagle 3.