[P] Realtime Stereo Vision (toy project)
Sup. I happened across one of the projects I was working on a few years ago, and I thought I would share it here in case you guys wanted to play with it.
tl;dr summary of the code is it does a shitty implementation of a few layers of lstm to process a live video feed from a webcam or two, and it tries to predict some configurable number of frames into the future, conditioned on its own action space also. Also, since I happen to have two identical webcams lying around, I adapted the code to also handle concurrent stereoscopic video feeds for the fun of it. It wasn’t really meant to be a super rigorous implementation (ie, I used lstm instead of cnn layers), but I still think it had some mildly interesting outcomes.
Anyways, here’s a link to the eye-bleedingly bad (but mostly functional) code: https://github.com/Miej/online-deep-learning