| | 1 | === Smart City |
| | 2 | |
| | 3 | Smart city is a demo application that shows how to deploy a real-time edge application on COMOS. This application takes the real-time video streaming as the input, running deep learning algorithms to detect object of interest, and putting masks upon them. |
| | 4 | |
| | 5 | The video camera is installed on Amsterdam Avenue, NY and delivers various resolutions at different frame rates. In this demo, we take H265 640 x 480 Constant rate 30 fps with 1024 Kbps. The backend object masking algorithm comes from the latest research work at Facebook https://github.com/facebookresearch/maskrcnn-benchmark. |
| | 6 | |
| | 7 | === Pre-set-up: |
| | 8 | 1. Create a reservation |
| | 9 | 1. Log into the console |
| | 10 | 1. Load image: {{{ omf load -t srv3-lg1.bed.cosmos-lab.org -i baseline.ndz -r 40}}} |
| | 11 | 1. Turn the node on: {{{ omf tell -a on -t srv3-lg1.bed.cosmos-lab.org }}} |
| | 12 | |
| | 13 | Next, we download a pre-configured container in order to minimize your experimental efforts. |
| | 14 | |
| | 15 | {{{docker push qingshanyouyou/smartcity:tagname}}} |
| | 16 | |
| | 17 | Go to the container. |
| | 18 | |
| | 19 | {{{nvidia-docker container run -it qingshanyouyou/smartcity:tagname bash}}} |
| | 20 | |
| | 21 | Go to the script folder. |
| | 22 | |
| | 23 | {{{cd /maskrcnn-benchmark/demo}}} |
| | 24 | |
| | 25 | Run the python script. |
| | 26 | |
| | 27 | {{{python smartcity.py}}} |
| | 28 | |
| | 29 | After running the python script, you will get the output video file named "output.avi". Note that we take a file to record the real-time processing result because of the slow forwarding speed from X11. |
| | 30 | |
| | 31 | Now, enjoy the video! |