Code Coming Soon!!
Load cells are attached to end effector of robots to measure the forces and moments while the end effector of the robotic arm is in contact with an object. Ideally, we should observe forces and moments only when there is an actual contact between the load cell and the object, but, we have observed that the inertial and gravitational load felt by the robot, while it is moving its arms without making contact with the object is significant. This induces errors in the force and torque values obtained from the sensors, making it difficult to trust the force and torque feedback we get from these sensors to build a closed loop controller for grasping. Therefore, the idea is to build a bi-directional LSTM based compensator, that can decouple the different forces experienced by the force/torque sensors, which will ideally compensate for the effect of gravity, inertia and drift on the load cell. This is being done using the data collected from the ATI-Mini45 force/torque sensor and uBot-6 at Laboratory of Perceptual Robotics, UMass Amherst.
Thermal images are mainly used to detect the presence of people at night or in bad lighting conditions, but perform poorly at daytime. To solve this problem, most state-of-the-art techniques employ a fusion network that uses features from paired thermal and color images. Instead, we propose to augment thermal images with their saliency maps, to serve as an attention mechanism for the pedestrian detector especially during daytime. We investigate how such an approach results in improved performance for pedestrian detection using only thermal images, eliminating the need for paired color images. For our experiments, we train the Faster R-CNN for pedestrian detection and report the added effect of saliency maps generated using static and deep methods (PiCA-Net and R3-Net). Our best performing model results in an absolute reduction of miss rate by 13.4% and 19.4% over the baseline in day and night images respectively. We also annotate and release pixel level masks of pedestrians on a subset of the KAIST Multispectral Pedestrian Detection dataset, which is a first publicly available dataset for salient pedestrian detection.
For people with severe motor disabilities, independent mobility is one of their biggest challenges. For people not capable of operating manual or remote controlled wheelchairs are heavily dependent on their attendants for all their basic needs. Therefore, there is a need for developing smart brain controlled wheelchairs which can help such people with basic movements without assistance. The project will generalise the use of EEG (electroencephalography) data for various multi user multi class classification tasks. Currently most EEG applications involve training and testing the classifier on the subject′s own data, which makes it hard for Brain Controlled devices to be ubiquitous. In order to do that, we used a Convolutional Neural Network on a 64 channel EEG Motor Imagery dataset from PhysioNet to train a Convolutional Neural Network (CNN) that takes in learned representations of the data which are built using an Auto-Encoder and obtained a classification accuracy of 75% when trained and tested on different dataset. We also run a user study using the Emotiv EPOC+ Headset to collect data under similar conditions as the training dataset, and test our network and obtain a classification accuracy of 65%
Link to the Dataset: PhysioNet Motor Imagery Dataset
Semantic Segmentation involves understanding the image on a pixel-by-pixel level i.e. to assign a class label to every pixel in the image. We experiment with different architectures to perform segmantic segmentation of images on the PASCAL VOC 2012 dataset. We implement the Fully Convolutional Networks (FCN) as our baseline method for performing semantic segmentation. We perform various experiments with the number and position of skip connections and adding different layers to aggregate more context information. We then implement an Improved Fully Convolutional Network (IFCN) architecture which introduces a context network that progressively expands the receptive fields of feature maps. In addition, dense skip connections are added so that the context network can be effectively optimized and fuses richscale context to make reliable predictions, which has proven to show significant improvements in segmentation on the PASCAL VOC 2012 dataset. We also modify the U-Net architecture for multi-class semantic segmentation with pre-trained weights from the VGG-16 architecture trained on the ImageNet dataset.
Link to the Dataset: PASCAL VOC-2012
Highlights in a sports video are the key exciting moments in the match which attract attention of the spectators in the match. A considerable amount effort is spent in extracting such highlights from the match requiring a lot of investment in terms of time and cost where the domain experts decide which frames must be included in the highlight, thus making it an expensive process, so there should be ways of generating automated highlights. We use audience reactions to classify frames in the game as ”highlights” and ”standard play”, because this method can be generalized to detect highlights in any sport. For this purpose, we have used the S-HOCK dataset which contains videos of spectators watching an ice-hockey match. In order to do so, we experiment with different methods to generate automatic highlights, using 3D Convolutional networks, HOG-SVM and pre-trained models on similar sports datasets. We obtained an accuracy of 26% using HOG-SVM, 67% using pre-trained models and 74% using 3D-CNN.
Link to the Dataset: S-HOCK Dataset
The public transportation system is a crucial part of the solution to the nation’s economic, energy and environmental challenges. This transportation system can be made more efficient, by studying the mobility of the users, from the sensor data collected by them. The current project have been developed to compare two methods of identifying the mode of transportation being used by a person using the location data collected from one’s cell phone.
The transportation modes, such as walking, driving, etc., that a user takes, can enrich the user’s mobility with informative knowledge and provide pervasive computing systems with more context information.
The raw data is time sequenced location data retrieved from the GPS in the cell phones of the users. This data was pre-processed in order to extract relevant features like velocity, acceleration and heading change and to filter out datasets with huge gaps and redundancy. A change point segmentation algorithm was applied where, the data was first broken down into segments using a loose bound of velocity and acceleration, segments smaller than a certain threshold were merged into the backward segment and consecutive segments of length smaller than another threshold were merged to form one segment was implemented using a Python Script. For every segment retrieved from the change point segmentation algorithm, the average velocity, the average acceleration and the average heading change has been calculated, and have been used as a feature for classification. The current work, uses selected trajectories of the GeoLife dataset, to train a decision tree based model, and uses GPS trajectory data collected from a user taking multimodal trips across Singapore, in order to test the model, using a Python script. A comparative study of the two algorithms i.e. Decision Tree and Random Forest was performed on some trajectories of the GeoLife Dataset and the Random Forest algorithm was found to be more accurate than the Decision Tree algorithm with an overall accuracy of 81.2%. Therefore, the change point segmentation and classification algorithms are highly suited to build a solution to the problem of discerning between motorised and non-motorised modes of transport, using a series of time-stamped GPS locations. In the real world, this project can not only help the land transport authorities make the public transportation system more efficient, but can also help the police in law enforcement, if a system which collects real time GPS logs of people is put in place. In order to achieve these results, Weka Machine Learning Toolkit, ArcGIS, MS Excel and Scikit Learn Library of Python is being used.
Link to the Dataset: GeoLife Dataset
The aim of this project was to develop a system that can assist drivers in finding a suitable parking spot in an open parking lot and can help him assess the traffic conditions on the road ahead of him. For this, I proposed an idea of developing a drone, which can take off and land from a moving car at the discretion of the driver, and send live video feed to the driver. The drone should identify the right car while landing and the battery used for the drone should be charged wirelessly from the car.
In a period of 3 months, I was able to successfully build an octocopter using the Pixhawk V4 flight controller. The drone has a flight time of about 20 min at the current weight and has a maximum payload capacity of 8kg. The drone is capable of taking off and landing autonomously from a surface and is GPS enabled. All the safety features provided by the Pixhawk V4 controller are enabled and tested. I have also designed a suitable landing gear for the drone using Plexi Glass. I also designed a circuit for a wireless charger for a single cell Lithium Polymer battery using EAGLE CAD.
This was done as an individual project as a part of an internship at the School of Vehicle Mechatronics, Techniche Universitat, Dresden, Germany
The aim of this project was to develop a system that can minimize road accidents. Our team of 3 designed a system that chalks out the major causes of road accidents and attempts to prevent them. The system has a driver drowsiness detection system, to check the state of the driver. This uses the Intel Real Sense camera to measure the percentage closure of eyes and to detect if the person is yawning. This is connected to the Intel NUC, which alerts the driver using audio stimulus, if he is found to be drowsy.
It also detects if the driver is under the influence of alcohol and if he has fastened the seat belt. The sensors for these are connected to the Atmel SAM D21 microcontroller which is in turn connected to the Intel NUC. If the driver is found to violate any rules, the NUC alerts the driver. The system also checks if the driver is tailgating or if somebody is tailgating from behind. If the driver is found to be tailgating repeatedly, the driver is warned and if needed, the details of the car are sent to the nearest police station using Wi-Fi for further action.
My role in this project was to realise face detection and eye detection using the Intel Real Sense camera and to integrate all the modules of the project using a C# program running on the Intel NUC.
Our team of 3 was selected among the top 10 teams of Atmel Embedded Design Challenge 2014 for this project. A research paper on this project has been selected for the 2nd International Conference on Contemporary Computing and Informatics, 2016.
The aim of this project was to devise a safety mechanism for treadmills that can prevent untimely deaths of people in fitness centres. Our team of 3 designed a prototype of a treadmill using 2 DC motors synchronised using a belt drive. The system designed asks for certain health parameters from the user, and sets the threshold for the maximum allowable heart rate. If the heart rate of the user exceed the threshold at some point in time during the course of the exercise, the speed of the treadmill is gradually reduced, the user is not allowed to increase the speed of the treadmill until his heart rate is below the threshold. My role in this project was to design the prototype of the treadmill and to program the controller to take in and process data from the user.
Built a device that enables the visually impaired to read. It consists of a camera, which scans the characters on any written material, converts it into its Braille equivalent using a MATLAB program and ejects the suitable dots on a panel placed, using servo motors. This enables the visually impaired to read whatever is placed under the reader.
My role in this project was to design a suitable actuation mechanism for the ejection of the braille dots on the panel.
Our team of 4, won the best project award at INK Makers Make-A-Thon, 2016 for this project.
The aim of this project was to build an amphibious robot that can help the victims of a natural calamity who are stuck in an inaccessible area. For this, our team of 3 designed and built a hovercraft capable of traversing autonomously to inaccessible locations and perform live environmental monitoring. This was performed by the Waspmote Gas Sensor board, which has inbuilt temperature, CO2 and O2 sensors and the XBee module. The sensor data captured from the environment is sent via the XBee module to the ground station.
The real time video feed from this vehicle to a ground station would assist the rescue team in planning and executing their operations efficiently. The live video feed was captured using a camera which was connected to the Intel Galileo board. This was transmitted the the ground station using Wi-Fi.
My primary role in this project was to program the Waspmote gas sensor board for capturing sensor data and transmitting using the Zig-Bee protocol.
This project was awarded as the best project at the Intel India Embedded Challenge in the category Internet of Things and Intelligent Systems.
The aim of this project was to build a cargo-aligning robot using the FireBird V robotics research platform. This platform utilizes microcontrollers Atmega 2560 and Atmega 8 which was programmed on AVR Studio 4 using Embedded C. The robot was programmed to traverse on a predefined path and the find the optimal route between its source and destination. Our team of 4 designed and built a robotic arm to pick up, turn and drop cargo blocks.
My role in this project was to design and implement algorithms to find the shortest path between the source and destination and to ensure that the robot does not deviate from the predefined path.
This project was a part of the e-Yantra National Robotics Competition conducted by Indian Institute of Technology, Bombay.
The aim of this project was to build a quadcopter in order to learn the basic controls of flying a quadcopter.I used NAZA M-Lite Flight Controller, a standard quadcopter chasis kit, brushless DC motors, electronic speed controllers, wireless camera and Tx-Rx kit.
This was an individual project done as a part of a summer internship program conducted by Indian Institute of Technology, Guwahati and Robosapiens Technologies, Delhi