ossdc / ossdc-visionbasedacc Goto Github PK
View Code? Open in Web Editor NEWDiscuss requirments and develop code for #1-mvp-vbacc MVP (see also this channel on ossdc.org Slack)
License: Apache License 2.0
Discuss requirments and develop code for #1-mvp-vbacc MVP (see also this channel on ossdc.org Slack)
License: Apache License 2.0
Collect datasets from similar situations and analyse them to help everyone understand what the car should have seen and what could have been done to avoid the accident.
Help us obtain the dataset from Uber SDC involved in the accident, at least 3 min before and 1 min after the impact (this is a reply to the police tweet with the video from the accident):
https://twitter.com/GTARobotics/status/976628350331518976
A few initial pointers to accident info:
The Google Maps StreetView link where the accident happened:
642 North Mill Avenue, Tempe, Arizona, USA
https://goo.gl/maps/wTDdCvSzc522
Brad Templeton analysis of the accident:
https://twitter.com/GTARobotics/status/976726328488710150
https://twitter.com/bradtem/status/978013912359555072
Experts Break Down the Self-Driving Uber Crash
https://twitter.com/GTARobotics/status/978025535807934470?s=09
Experts view on the fact that LIDAR should have detected the person from far away:
https://twitter.com/GTARobotics/status/977764787328356352
This is the moment when we decide that human lives matter more than cars
https://www.curbed.com/transportation/2018/3/20/17142090/uber-fatal-crash-driverless-pedestrian-safety
Uber self-driving system should have spotted woman, experts say
http://www.cbc.ca/beta/news/world/uber-self-driving-accident-video-1.4587439
IIHS shows the Volvo XC90 with a range just under 250 feet (76 meters) with "low beams" on!
https://twitter.com/GTARobotics/status/977995274122682368
Help us get current companies that test SDC to provide datasets from their own cars in similar situations as the accident:
https://twitter.com/GTARobotics/status/977773180344512512
Lets also capture current SDC sensors configurations/specs in:
https://github.com/OSSDC/OSSDC-Hacking-Book/wiki
Join the discussions on OSSDC Slack at http://ossdc.org
Objects Detection and Tracking Using Points Cloud Reconstructed from Linear Stereo Vision
https://www.intechopen.com/books/current-advancements-in-stereo-vision/objects-detection-and-tracking-using-points-cloud-reconstructed-from-linear-stereo-vision
See more articles in this area:
One-Shot Video Object Segmentation.
Use https://github.com/zdzhaoyong/GSLAM for VisionBasedACC.
Integrate in Google Colaboratory.
Collect YouTube, Vimeo etc videos links, with all kinds of driving, that are good for testing SDC algorithms especially in the area of LKAS, VisionBased ACC, FCW, Pedestrian avoidance, etc.
You can try a few algorithms here, for free in the cloud with GPU acceleration, more will come soon:
Here is my list of public videos as a starting point:
https://www.youtube.com/playlist?list=PLUop7b1Q1uZkv5__d2yPZG1cAXcelata8
Add your contributions in the comments here on this issue, or as comments on this video:
https://youtu.be/Z3bxoi0ZJ_g
NoScope: 1000x Faster Deep Learning Queries over Video
https://dawn.cs.stanford.edu/2017/06/22/noscope/
Code:
https://github.com/stanford-futuredata/noscope
And test it on the videos here:
#19
Geometric Deep Learning | Michael Bronstein || Radcliffe Institute
https://youtu.be/ptcBmEHDWds
Hey,
I watched your video here: https://www.youtube.com/watch?v=_cEouvyJNMs
And it led me to your python notebook that contains ssd net for object detection. I want to try out the distance estimation thing which isn't there in the notebook. Can you please link me with the correct notebook if this one's wrong or something else?
Thanks
SurfNet: Generating 3D shape surfaces using deep residual networks
Here is a dataset collected from two (not perfectly aligned) PS3 Eye cameras:
20180209_1736 - 2 x PS3 Eye + Android Sensors + OBD II data
https://drive.google.com/open?id=1KBWIUBp5nZNDiNcKVJP1Tzqs_5ll5S4Y
Left camera (3 GB)
https://drive.google.com/open?id=1WAv2m-APu0_rjMptRfK2DxRdvu2eWlvc
Right camera (3 GB)
https://drive.google.com/open?id=1-LMnWK2L_CRQMMBhGu9575HHhST2HLHP
Video with processed sample data from left camera (20180209_1736/left/001%4d.jpg):
https://www.youtube.com/watch?v=_cEouvyJNMs
The dataset contains the images from left and right cameras, sensors data (GPS, Accelerometer, Gyro, Compass, Magnetometer from Android phone) collected with Android Sensors app and OBD II data collected with Torque Lite Android app.
The images needs to be aligned and synchronized (can be done by using the signal or break lights of the cars) also with the sensors and OBD II data (based on start/stop periods).
Both the alignment and synchronization should be done automatically, so a program to do this will be required, to ease the use of future datasets like this
Machine Learning Meets Geometry
https://youtu.be/pgomVDoEMrc
Hello, I executed your Jupyter notebook in Colaboratory, environment is Python3 and GPU enabled.
I got this error on cell [12]:
ModuleNotFoundError: No module named 'youtube_dl'
To solve it please add before cell [12]:
os.environ['PAFY_BACKEND'] = 'internal'
Project code:
Robust Stereo Visual Inertial Odometry for Fast Autonomous Flight
https://github.com/KumarRobotics/msckf_vio
Amazing demo video here:
https://youtu.be/jxfJFgzmNSw
Karl Granstrom, Marcus Baum, Stephan Reuter
This article provides an elaborate overview of current research in extended object tracking. We provide a clear definition of the extended object tracking problem and discuss its delimitation to other types of object tracking. Next, different aspects of extended object modelling are extensively discussed. Subsequently, we give a tutorial introduction to two basic and well used extended object tracking approaches - the random matrix approach and the Kalman filter-based approach for star-convex shapes. The next part treats the tracking of multiple extended objects and elaborates how the large number of feasible association hypotheses can be tackled using both Random Finite Set (RFS) and Non-RFS multi-object trackers. The article concludes with a summary of current applications, where four example applications involving camera, X-band radar, light detection and ranging (lidar), red-green-blue-depth (RGB-D) sensors are highlighted.
Mask R-CNN in TensorFlow.
This repo attempts to reproduce this amazing work by Kaiming He et al. :
Mobileye YouTube video
Try to reproduce this behavior in Vision Based ACC project.
Collect test videos from IIHS and other ADAS testing organizations, and also contribute with your own tests.
Provide in depth analysis of each video and situation.
See these kinds of articles, Mobileye specific:
https://scholar.google.ca/scholar?q=mobileye+vision+based+acc+paper
BoxCars: Improving Fine-Grained Recognition of Vehicles using 3D Bounding Boxes in Traffic Surveillance.
A declarative, efficient, and flexible JavaScript library for building user interfaces.
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
An Open Source Machine Learning Framework for Everyone
The Web framework for perfectionists with deadlines.
A PHP framework for web artisans
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
Some thing interesting about web. New door for the world.
A server is a program made to process requests and deliver data to clients.
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
Some thing interesting about visualization, use data art
Some thing interesting about game, make everyone happy.
We are working to build community through open source technology. NB: members must have two-factor auth.
Open source projects and samples from Microsoft.
Google ❤️ Open Source for everyone.
Alibaba Open Source for everyone
Data-Driven Documents codes.
China tencent open source team.