March 27, 2013

Pause 4 Relax : Software Review



PAUSE 4 RELAX

Pause 4 Relax is a handy and reliable application designed to help one to relax his eyes to avoid eye-sight problems.

Regular computer users who are exposed to the heat waves from the monitor are affected with the eye-sight problems. Hence it is recommended to give rest to the eyes for 5 min every half-an-hour. This software paves a way for the users by remembering every 25 min with a notification. The brightness of the screen dims for 5 min and restores thereafter. This sequence continues till shutting down the computer.

One can skip a relaxation for one time and can also switch it off if one is playing games or watching movies.


Features

  • Relax your eyes every 30 mins [still you can manage your own timing]
  • Skip relaxation for one time and add 5,10,15 min to the running time
  • Option to disable relaxation during playing games or watching movies
  • Take relaxation even before the actual relaxation time begins
  • Set notification sounds if you want
  • Show / Hide the "Skip Now" button in Relaxation window
  • Simple and very light software
  • Progress bar in the relaxation window to show the time remaining pictorially.
  • No need to install - double click to execute
  • Portable
  • Option to load at start with windows
  • Minimal software requirements and memory



System Requirements

  • No specific requirements necessary
  • Runs on all windows machine with .Net framework


Screen shots:







Certification


Download Links
   

     

January 1, 2013

intelliPaper reveals disposable, paper-based USB drives


USB drives have become so prolific in recent years that they've become practically disposable. Now, one company has created a different type of flash drive that can literally be crumpled up and thrown in the garbage. With an embedded silicon chip, intelliPaper seamlessly turns an ordinary strip of paper into a fully functioning USB drive.

We've seen concepts for a USB stick made out of paper before, but the design team at intelliPaper has now patented technology that allows electronic components to be layered into a regular sheet of paper with USB contact points exposed. Once it's ripped from the full sheet and folded in half, the paper can be inserted into any USB port to access the files it holds, just like a typical USB drive. Files can be added and removed like any other storage device, and the drive can be reused for as long as the paper and contacts remain intact.

The paper used is about as thick as card stock, and the embedded chip can hold 8-32 MB of data – the team has not decided on a fixed capacity yet, but it will be within this range. So far, the developers have implemented their technology in mail-out flyers, promotional brochures, and business cards, among other items. Uploading data to a fresh card does require a special reader and some software to avoid damaging it, which intelliPaper offers to customers. If someone doesn't want to risk damaging the paper drive itself, intelliPaper also communicates wirelessly with any near field-enabled smartphone or tablet.

Once it's ripped from the full sheet and folded in half, the paper can be inserted into an...


Since intelliPaper is so inexpensive to produce, it's not hard to think of plenty of uses for such a paper-based USB drive. Tourists could send out postcards with vacation photos uploaded onto them, couples could send wedding invites with a digital version attached, and schools and businesses could hand out multiple pages of documents uploaded to a single sheet of paper.

The company mostly ships intelliPaper to customers in bulk, but it is also preparing to release products aimed at individual consumers (greeting cards, note cards, and so on). The design group recently finished an Indiegogo funding campaign that was intended to speed up this process, but only raised US$6,480 of its $300,000 goal. Nevertheless, the team did manage to find a U.S. distributor and plans to release USB-enabled note cards, called "DataNotes," in mid-2013.

Check out the video below to see intelliPaper president/CEO Andrew DePaula demonstrate some of the many uses for a sheet of paper outfitted with USB capabilities.

Source: intelliPaper

December 29, 2012

Roboy team aims to build robot toddler in nine months


If robots are going to be part of our everyday lives, they’ll need to fit into our homes rather than the factory floor. Few people would be comfortable living with a metal spider on tank treads, so the University of Zurich’s Artificial Intelligence Laboratory (AI Lab) is building a robot toddler called “Roboy.” Using “soft robotics” technology that mimics the human body, the 1.2 meter (3 ft, 11 in) tall humanoid robot is part of an effort to make robots that people are more comfortable with in day-to-day situations.

Roboy doesn't look very endearing at the moment. In fact, it looks more like a cyborg skeleton than a charming child, but it’s still a work in progress. The laboratory’s goal is to build Roboy in only nine months. Work began last June with 15 project partners and over 40 engineers and scientists. These parties are providing expertise and funds through sponsorship and crowdfunding that includes auctioning space on the robot for logos, and hiring it out for business functions when completed.

Roboy's features

Roboy is based on the laboratory’s previous project, the humanoid, frighteningly cycloptic Eccerobot. Built out of plastic, Roboy is modeled on the human musculoskeletal system, but this mimicry goes beyond the aesthetic. Instead of motors in its joints, Roboy uses motor assemblies that pull elastic cables, so the system operates in a way similar to muscles and tendons. AI Lab claims that this will allow Roboy to move “almost as elegantly as a human.”

Currently, Roboy is more of a research project than an engineering enterprise. The team is developing new technologies with an eye toward scalable production using CAD and 3D printing to allow for full production of robots within days of development.

The purpose of Roboy is to push for the acceptance of service robots by making people more comfortable having them around all the time. With an aging population, AI Lab believes that such service robots will be increasingly important in helping the elderly to continue independent lives.

Roboy is currently getting a new face chosen by a Facebook contest, and can move its arms. Later, the robot will be covered with a soft skin. Roboy will make its first public appearance at the “Robots on Tour” exhibition on March 9, to celebrate AI Lab’s 25th anniversary.

The video below shows Roboy in action.

Source: Roboy, AI Lab


December 10, 2012

LuminAR Bulb transforms any surface into a touch screen

The LuminAR Bulb works with standard light fixtures and projects an interactive image onto...

We've all seen gigantic touch screens on the news or in movies, but what if you could achieve the same type of interface by simply replacing the bulb in your desk lamp? That's the idea behind the LuminAR, developed by a team led by Natan Linder at the MIT Media Lab's Fluid Interfaces Group. It combines a Pico-projector, camera, and wireless computer to project interactive images onto any surface – and is small enough to screw into a standard light fixture.

The LuminAR project (the capitals reflect its shared properties with other augmented reality set-ups) has two separate but interconnected components. The Luminar Bulb is a stand-alone unit that allows users to interact with its projection through simple hand gestures for zoom, position control, and content manipulation. It can plug into any fixture, but takes on even more functionality when combined with the LuminAR Lamp - an articulated robotic arm (similar to the Pinokio Lamp), enabling you to move the projected image around.

The Luminar Lamp remembers where you've moved different applications, allowing you to organize your workspace accordingly, such as putting your twitter feed in a less distracting location, or projecting a Skype session onto a wall. The Lamp can also take snapshots of the work area, allowing you to quickly scan and share work documents seamlessly across multiple devices.

Besides tracking your hands and fingers, the camera and image processing software could detect objects in the work space, such as a canned soft drink, and automatically display targeted advertising around it. One potential application would be projecting rich media, including product information, in a retail setting. In effect, browsing a store's display could incorporate the same media and interactivity as a product web site.

The LuminAR project was developed through 2010, and showcased earlier this year at the ACM CHI Conference on Human Factors in Computing Systems.

See how it works in this video summarizing its development.


Source: Fluid Interfaces Group




November 22, 2012

Breakthrough 3D gesture controller opens new vistas for human-computer interfacing


The smallest gesture can hide a world of meaning. A particular flick of a baton and a beseeching gesture can transform the key moment of a concert from mundane to ethereal. Alas, computers are seriously handicapped in understanding human gestural language, both in software and hardware. In particular, finding a method for describing gestures presented to a computer as input data for further processing has proven a difficult problem. In response, Microchip Technologies has developed the world's first 3D gesture recognition chip that senses the gesture without contact, through its effect on electric fields.
The MGC3130 chip in a 5x5mm package, resting on a fingertip
The MGC3130 chip in a 5x5mm package, resting on a fingertip
Microchip Technologies has recently unveiled their GestIC technology as implemented in the soon-to-be-available MGC3130 chip, an outgrowth of anearlier technology. When used as a 3D digitizer, the MGC3130 resolves position within a 15 cm (6 in) cube at a remarkable resolution of 150 dpi. (Yes, that's vertical resolution as well as in the plane, meaning that roughly a billion voxels (3D pixels) can be distinguished within the scanning volume.) The sampling rate is 200 measurements per second, allowing the GestIC technology to follow quick adjustments of hand and finger positions, velocities, and accelerations.
The MGC3130 enables a new approach to the problem of human-machine interfacing (HIM), recognizing gestures by measuring the changes in an electric field as the gesture is made. When gestures are sensed via their effect on electric fields, the step of precisely measuring hundreds of positions for each millisecond of a gesture and converting that data into a concise description of a gesture is no longer needed. Instead, a vastly simpler procedure can be adopted. The output of an electric field-based gesture sensor is itself something of a gestalt of a gesture, which has the potential to greatly simplify the interpretation of gestures.
The electric field lines produced by the GestIC technology electrodes in the absence of an...
The electric field lines produced by the GestIC technology electrodes in the absence of any perturbing influence
GestIC technology detects gestures through the changes which appear in a circumambient electric field. The chip generates an excitation voltage having a frequency around 100 kHz. The excitation voltage is applied between the transmitter electrode and a ground plane (in commercial practice, the function of the ground plane will be taken by the device using GestIC technology). This sets up an electric field that extends from the transmitter electrode into the scanning region above the electrodes. As the wavelength of the excitation voltage is far larger than the size of the electrodes, the electric field is quite uniform through out the scanning region.
The same electric field adjusts to the presence of a hand
The same electric field adjusts to the presence of a hand
When a user reaches into the scanning area, the electric field changes in response. Electric field lines must approach a conducting body perpendicular to the surface of that body. This is shown in the image above, where the field lines which pass near the hand are shunted to ground through the conductivity of the human body itself. (The person operating the device must be grounded to the ground plane.) The position of the hand within the sensing volume causes a compression of the equipotential lines and reduces electrode signal levels.
Block diagram of the MGC3130 chip driving a set of sensing electrodes
Block diagram of the MGC3130 chip driving a set of sensing electrodes
Instead of producing a scanned map of points on the surface of the hand, however, the MGC3130 measures a small number of analog voltages – the five voltage differences between the various electrodes and the ground plane. This analog data provides a highly compressed signature of the gesture. It can't be used to uniquely model the position of the hand, as there is not enough redundancy in the data. Despite this, this data can be used to accurately identify gestures.
A given gesture always produces the same signature, and gestures close to the given gesture will produce similar signatures, as will the same gesture being presented by a user with a larger or smaller hand. (This is equivalent to saying that mapping electric field gesture detection onto actual gestures is mathematically a continuous function.) As the system is now dealing with tens of bits of data instead of thousands of bits of data, the job of recognizing patterns associated with particular types of gestures becomes far easier, in analogy to the image preprocessing which occurs in the retina before the processed data is presented to the visual cortex.
Electrode geometry for Microchip's GestIC gesture recognition technology
Electrode geometry for Microchip's GestIC gesture recognition technology
Imagine that a user places their hand in the sensing volume, and then curls their thumb and forefinger together in an "A-OK" gesture. The GestIC sensor produces five voltages which are characteristic of that gesture. It doesn't know or care (speaking anthropomorphically) that the thumb and forefinger are touching and the other three fingers are splayed outward. Neither can a computer determine the position of the hand by analysis of the five voltages: the detailed position information is simply not in that data.
Instead, the sensor's MPU says to itself "the voltages swooped about pretty quickly with time and then settled down into a new pattern. I guess this is a new gesture. Let's compare the sizes of the present voltages with a bunch of patterns of standard gestures in my memory. Hmmmm. These voltages seem to match pretty well with a slightly rotated "A-OK" gesture – at least, better than anything else in my recognition patterns. Don't know what that means, but I'll send my decision over to be used as input data by the rest of the program."
The layers of programming support provided by Microchip for their GestIC system
The operational software that emulates this inner dialog is part of the Colibri software suite that supports the chip. Comparison and recognition of input patterns is carried out by a stochastic Hidden Markov model analysis that is preprogrammed with a reliable set of standard 3D hand and finger gestures (no, not that one) that can be easily employed in their products. Examples include position tracking (essentially digitizing the position of a fingertip), flick, circle, and symbol gestures, and many more. A system can also be activated from a standby condition by a stylized gesture. The Colibri suite allows developers to rapidly and inexpensively incorporate GestIC technology into products, as the programming for a basic human-machine interface has been provided.
At present, the MGC3130 chip will be supplied in the 28 lead 5x5-mm QFN form factor. The frequency of the electric field is variable between 70 and 130 kHz, and the firmware on the 3130 chip enables frequency hopping to substantially eliminate RF interference. The power requirements are very small, about 100 milliwatts while actively detecting and processing gestures, about 150 microwatts in standby mode, and about 30 microwatts in a deep sleep mode. Both an MPU processor and a firmware version of the Colibri software suite are integrated on the one chip.
Sample circuitry for an operational 3D gesture recognition system built around the MGC3130...
Sample circuitry for an operational 3D gesture recognition system built around the MGC3130 chip
Only a set of electrodes and eleven discrete electronic components are required for full operation of a GestIC system. This circuitry is provided in Microchip’s Sabrewing MGC3130 Single Zone Evaluation Kit. The Sabrewing comes with selectable 5" or 7" electrodes and the AUREA Graphical User Interface, which allows designers to easily match their system commands to Microchip’s Colibri Suite (also included). The evalutation kit costs US$169.
Let's imagine an application perhaps two generations down this development path. In front of you appears a somewhat larger set of electrodes. The transmitter electrode delivers an electric field of two different frequencies, while your right and left arms are connected to the ground plane through filters. In this way, gestures of your right hand can be separated from gestures of your left hand. The application is a 3D sculpturing program, in which virtual clay is formed by the motions of your hands and fingers. The virtual clay could be spinning for throwing pots, or fixed for more traditional sculpture. Once you obtained a pleasing sculpture, the program would send the description to a 3D printer or CAM system to fix it in the sculpting material of your choice.
Such machines might be available not only for professional sculpture, but at a suitable price point might be used to encourage artistic talent and imagination in children. Who knows, perhaps even artists of other species (elephants, apes, etc.) might benefit from this new technology…
The video below presents a panoramic overview of the GestIC system.