Machine Vision Development Module
Language, Tools and Environment :
C, C++, C# .NET, LabVIEW, Visual Studio, Windows, Linux, VxWorks.
Description :
Machine Vision development module comprises of distinctive and intricate image processing algorithms to build "Machine Vision" systems.
These machine vision systems are increasingly being used in "Vision Guided Robotics", "Automotive and Aerospace" , "Medical and Scientific", "Semiconductor and Solar" , "Food, pharmaceuticals and consumer goods" and multiple other domains.
I am involved in the development of algorithms for "Optical Character Recognition" (OCR) for industrial robotics and inspection systems. The OCR module implements intelligent algorithms for "Thresholding", "Segmentation" and "Classification". I am also involved in the task of Platform portability and Multicoring of Vision Algorithms on various real time architectures.
These machine vision systems are increasingly being used in "Vision Guided Robotics", "Automotive and Aerospace" , "Medical and Scientific", "Semiconductor and Solar" , "Food, pharmaceuticals and consumer goods" and multiple other domains.
I am involved in the development of algorithms for "Optical Character Recognition" (OCR) for industrial robotics and inspection systems. The OCR module implements intelligent algorithms for "Thresholding", "Segmentation" and "Classification". I am also involved in the task of Platform portability and Multicoring of Vision Algorithms on various real time architectures.
Intel Perceptual Computing - Comprehensive Gesture, Gaze and Speech Recognition Framework to replace conventional Modes of Human Computer Interaction.
Language, Tools and Environment :
Intel's "Interactive Gesture Camera" and "Perceptual Computing SDK"
Description :
Primary goal of this research work, is to realize distant vision of interacting with computers and other systems in 3D space without the use of any physical interaction devices like Keyboard, Mouse, Joystick etc.
The prevalent conventional modes of interaction shall be substituted by highly immersive and intuitive interaction modes like "Gestures, Gaze and Speech Recognition".
These natural modes of interactions make an individual to feel working with computers as simple as raising a hand, speaking out a word, looking around etc. And intricacies of working with computers is highly abstracted through these interaction means.
The framework also enables users from all the age groups to work with computers with or without pertinent knowledge of working with such systems.
Realizing the above conceptualized framework needs a robust hardware, which can most accurately track and perceive human gestures and synthesize voice into appropriate commands.
And a software platform of highly efficient algorithms and APIs, to best utilize the hardware.
Intel's "Interactive Gesture Camera" and "Perceptual Computing SDK" are the right combination of a "Robust Hardware and a comprehensive SDK" in achieving the above conceptualized framework.
Use Case :
The prototype will be primarily used for , interacting with desktop based applications, which are commonly used in day to day work.
For example -
Gesture Based Interactions- Browsing the web, Drawing on a canvas, Slide shows, Start or Exit applications, Minimize, Maximize windows.
Voice Based Interactions- Creating or Editing a document by voice to text synthesis. Voice based user identification for security features
Gaze Based Interactions- Mouse tracking, Emotion analysis.
Advanced Use Cases
Gaming, Security Systems, ease of interaction
The prevalent conventional modes of interaction shall be substituted by highly immersive and intuitive interaction modes like "Gestures, Gaze and Speech Recognition".
These natural modes of interactions make an individual to feel working with computers as simple as raising a hand, speaking out a word, looking around etc. And intricacies of working with computers is highly abstracted through these interaction means.
The framework also enables users from all the age groups to work with computers with or without pertinent knowledge of working with such systems.
Realizing the above conceptualized framework needs a robust hardware, which can most accurately track and perceive human gestures and synthesize voice into appropriate commands.
And a software platform of highly efficient algorithms and APIs, to best utilize the hardware.
Intel's "Interactive Gesture Camera" and "Perceptual Computing SDK" are the right combination of a "Robust Hardware and a comprehensive SDK" in achieving the above conceptualized framework.
Use Case :
The prototype will be primarily used for , interacting with desktop based applications, which are commonly used in day to day work.
For example -
Gesture Based Interactions- Browsing the web, Drawing on a canvas, Slide shows, Start or Exit applications, Minimize, Maximize windows.
Voice Based Interactions- Creating or Editing a document by voice to text synthesis. Voice based user identification for security features
Gaze Based Interactions- Mouse tracking, Emotion analysis.
Advanced Use Cases
Gaming, Security Systems, ease of interaction
Embedded Computer Vision Framework on Pandaboard, and Gesture Recognition
Language, Tools and Environment :
C++, OpenCV, Ubuntu.
Description :
This is an ongoing research venture solely undertaken by me, to leverage the multifarious functionalities provided by OpenCV, on tiny but yet robust hardware platform "Pandaboard" running Ubuntu. The main aim of the project is to build a gesture recognition framework, based on appropriate image processing algorithms provided by OpenCV.
The advantage of this product is that, since the gesture recognition framework runs on Pandaboard, the whole system is physically portable without any hassles, and it can be integrated with the any of the target device/platform with only subtle modifications. This project enables various solutions for "Vision Based" systems, which primarily run on non static targets like robots.
The advantage of this product is that, since the gesture recognition framework runs on Pandaboard, the whole system is physically portable without any hassles, and it can be integrated with the any of the target device/platform with only subtle modifications. This project enables various solutions for "Vision Based" systems, which primarily run on non static targets like robots.
Measure Data Analyzer (MDA)
Language, Tools and Environment :
VC++, Borland C++, VCL, COM Programming on Window 2000/XP and Vista
Description :
"Measured Data Analyzer" (MDA), is calibration tool which is primarily used to tune the engines for best performance by post processing the recorded data from an online calibration tool called "INCA" ( Integrated Calibration and Acquisition System). MDA and INCA, product suite was most crucial in calibrating the efficiency, performance, thrust and throughput of the engines. Working on this product I implemented different ways to visualize every minute data, and developed algorithms to interpolate and extrapolate the missing data. To enable calibration engineers to share and discuss their analysis, I developed an intuitive print layout editor, which could generate the output of analysis in the form of vectorized images, which maintain integrity even at the highest zoom level.
FlashLite 4.x plugin for Qt Browser on Qt Webkit
Language, Tools and Environment :
Qt Framework, Qt C++, Symbian C++, Carbide.c++, SVN, Rational Synergy and Rational Change. Symbian^3
Description :
This project involves, the development of Qt FlashLite Plugin for Nokia Qt Browser on Qt Webkit which enables video playback from Webvideo sites like YouTube, Google videos etc.
Being part of the development team I was involved in the research and development of "Adobe Flash Player" for Symbian platform. Working closely with "Adobe" I implemented Qt Flash plugin for browsers, from an idea to a prototype, and to the complete product. During this work I worked extensively on the development of Gesture Recognition module, and implemented an algorithm to improve web page scrolling performance by exchanging bitmaps with browser rendering engine and switching between software and hardware rendering modes appropriately
Being part of the development team I was involved in the research and development of "Adobe Flash Player" for Symbian platform. Working closely with "Adobe" I implemented Qt Flash plugin for browsers, from an idea to a prototype, and to the complete product. During this work I worked extensively on the development of Gesture Recognition module, and implemented an algorithm to improve web page scrolling performance by exchanging bitmaps with browser rendering engine and switching between software and hardware rendering modes appropriately
Game Studio
Language, Tools and Environment :
C# .NET programming on Window 2000/XP
Description :
"Game Studio", is an application which generates huge 3D canvas for games which run on Set-Top box. On server side the application converts 3D canvas into byte array representing the indexes of different image pallets and Broadcasts this data to clients running on multiple Set-Top boxes. All the clients will then reconstruct the 3D canvas by using the received image pallets and index array.
iTrack
Language, Tools and Environment :
C# , .Net, XML persistence, Windows XP/Vista/7
Description :
iTrack is a tool to maintain a record of day to day activities of an employee. A user can create TO-DO list, plan future engagements, review past work history. The tool provides integration with e-mail clients. iTrack also provides a quick notes application.