In 2012, TurtleBot2 was developed by Yujin Robot based on the research robot, iClebo Kobuki. It was developed in 2010 and has been on sale since 2011. TurtleBot1 was developed by Tully (Platform Manager at Open Robotics) and Melonee (CEO of Fetch Robotics) from Willow Garage on top of the iRobot’s Roomba-based research robot, Create, for ROS deployment. There are 3 versions of the TurtleBot model. Since then TurtleBot has become the standard platform of ROS, which is the most popular platform among developers and students. TurtleBot, which originated from the Turtle of Logo, is designed to easily teach people who are new to ROS through TurtleBot as well as to teach computer programming language using Logo.
The nine dots used in the ROS logo derived from the back shell of the turtle. It is also used to create the Turtle icon as a symbol of ROS. In addition, the turtlesim node, which first appears in the basic tutorial of ROS, is a program that mimics the command system of the Logo turtle program. Turtle is derived from the Turtle robot, which was driven by the educational computer programming language Logo in 1967. TurtleBot is a ROS standard platform robot. The Jetson Nano Developer Kit setup must be completed first. Please refer to the video below in order to set up the Jetson Nano for TurtleBot3. TurtleBot3 Hardware is compatible with Jetson Nano SBC. Projects developed by third-party developers. If you use our work in your project, we would love you to include an acknowledgement and fill out our survey. This work is licensed under the MIT License.
Ira Kemelmacher-Shlizerman, University of Washington. Brian Curless, University of Washington. Soumyadip Sengupta, University of Washington. Andrey Ryabtsev*, University of Washington. Shanchuan Lin*, University of Washington. The original paper uses train_base.pth to train only the base model till convergence then use train_refine.pth to train the entire network end-to-end. For detail about using our model, please check out the Usage / Documentation page.Ĭonfigure data_path.pth to point to your dataset. You can run our model using PyTorch, TorchScript, TensorFlow, and ONNX. The script only works on Linux system and can be used in Zoom meetings. We provide a demo application that pipes webcam video through our model and outputs to a virtual camera. inference_webcam.py: An interactive matting demo using your webcam.Īdditionally, you can try our notebooks in Google Colab for performing matting on images and videos. inference_video.py: Perform matting on a video. inference_images.py: Perform matting on a directory of images. More detailed instructions are included in the files. We provide several scripts in this repo for you to experiment with our model. HD videos (by Sengupta et al.) (Our model is more robust on HD footage). We updated our project to MIT License, which permits commercial use. PhotoMatte85 dataset is now published. VideoMatte240K dataset is now published. Paper received CVPR 2021 Best Student Paper Honorable Mention. For more architecture detail, please refer to our paper.Ĭheck out Robust Video Matting! Our new method does not require pre-captured backgrounds, and can inference at even faster speed! For production use, you are expected to do additional engineering for hardware encoding/decoding and loading frames to GPU in parallel. The inference_video.py script allows you to test your video on our model, but the video encoding and decoding is done without hardware acceleration and parallization. The inference_speed_test.py script allows you to measure the tensor throughput of our model, which should achieve real-time. Our research's main contribution is the neural architecture for high resolution refinement and the new matting datasets. Our model requires capturing an additional background image and produces state-of-the-art matting results at 4K 30fps and HD 60fps on an Nvidia RTX 2080 TI GPU.ĭisclaimer: The video conversion script in this repo is not meant be real-time. Official repository for the paper Real-Time High-Resolution Background Matting. Real-Time High-Resolution Background Matting