Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[P] Python library to work with the Visual Wake Words Dataset.

[P] Python library to work with the Visual Wake Words Dataset.

Recently Google published a paper introducing the Visual Wake Words Dataset.

Currently vision models are benchmarked on the CIFAR10 or ImageNet datasets both of which are restricted in terms of benchmarking the model accuracy and the memory costs for the common low-complexity microcontroller use-case. We present a new dataset, Visual Wake Words, that represents a common microcontroller vision use-case of identifying whether a person is present in the image or not, The proposed dataset is derived from the publicly available COCO dataset, and provides a realistic benchmark for tiny vision models.

As the dataset is derived from the COCO dataset I created a library that inherits from the pycocotools libary and that can be used in a similar fashion on the Visual Wake Words Dataset.

I’ve also included a Pytorch Dataset class that can be used like any VisionDataset.

https://github.com/Mxbonn/visualwakewords

https://i.redd.it/xxeoua9xp2931.png

https://i.redd.it/3lyaebaxp2931.png

submitted by /u/Mxbonn
[link] [comments]