{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"compute-engine","owner":"larq","isFork":false,"description":"Highly optimized inference engine for Binarized Neural Networks","allTopics":["android","tensorflow","keras","simd","armv7","aarch64","armv8","aarch32","tflite","mlir","raspberry-pi","inference","bnn","binarized-neural-networks","larq"],"primaryLanguage":{"name":"C++","color":"#f34b7d"},"pullRequestCount":2,"issueCount":16,"starsCount":240,"forksCount":33,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-24T11:23:35.150Z"}}],"repositoryCount":1,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}