New top story on Hacker News: Show HN: Zero-power photonic language model–code
Show HN: Zero-power photonic language model–code
5 by damir00 | 2 comments The model uses a 1024-dimensional complex Hilbert space with 32 layers of programmable Mach–Zehnder meshes (Reck architecture) and derives token probabilities directly via the Born rule. Despite using only unitary operations and no attention mechanism, a 1024×32 model achieves coherent TinyStories generation after < 1.8 hours of training on a single consumer GPU. This is Part 1 - the next step is physical implementation with $50 of optics from AliExpress. on Hacker News.
The model uses a 1024-dimensional complex Hilbert space with 32 layers of programmable Mach–Zehnder meshes (Reck architecture) and derives token probabilities directly via the Born rule. Despite using only unitary operations and no attention mechanism, a 1024×32 model achieves coherent TinyStories generation after < 1.8 hours of training on a single consumer GPU. This is Part 1 - the next step is physical implementation with $50 of optics from AliExpress.
5 by damir00 | 2 comments The model uses a 1024-dimensional complex Hilbert space with 32 layers of programmable Mach–Zehnder meshes (Reck architecture) and derives token probabilities directly via the Born rule. Despite using only unitary operations and no attention mechanism, a 1024×32 model achieves coherent TinyStories generation after < 1.8 hours of training on a single consumer GPU. This is Part 1 - the next step is physical implementation with $50 of optics from AliExpress. on Hacker News.
The model uses a 1024-dimensional complex Hilbert space with 32 layers of programmable Mach–Zehnder meshes (Reck architecture) and derives token probabilities directly via the Born rule. Despite using only unitary operations and no attention mechanism, a 1024×32 model achieves coherent TinyStories generation after < 1.8 hours of training on a single consumer GPU. This is Part 1 - the next step is physical implementation with $50 of optics from AliExpress.
Comments
Post a Comment