Posts with «arduino micro» label

Unlock your garage door with a fingerprint

Using an Arduino Micro for control, French teenager “Joebarteam” came up with a way to biometrically secure his garage.

If you need to get into your locked garage, what could be better than using your fingerprint? To this end, Joe’ came up with a system that unlocks his door using a fingerprint scanner, and a bistable relay to disable communication between the scanner and the Arduino if there’s a problem.

It’s a really professional-looking build, and the locking mechanism is especially interesting. Two rack-and-pinion devices plunge shafts into the ground, making the door impossible to open (it has to pull out before going up). If there is an issue with the system, the pins can be physically unlocked and disabled as needed.

You can find more details on the project’s Instructables page here.

An interactive Lea shapes puzzle for visually impaired children

Using two Arduino Micros, these parents created a unique Lea shapes puzzle for their daughter Rebecca.

Because of a medical condition, Rebecca will need to have her vision assessed at the age of two or younger. This means that she’ll have to be tested without yet knowing her alphabet, and instead need to be familiar with Lea shapes which can act as a substitute for letters in a vision test. Unfortunately, her hearing is also impaired, meaning that a non-visual type of stimulus is needed to encourage a correct response.

To accomplish this, her parents came up with an excellent puzzle system, where when one of these shapes is dropped into the correct slot, the smart lights in the room change to the corresponding color. It’s an interesting project that will hopefully help with a pressing need.

In terms of hardware, a Raspberry Pi 3 is used as a central hub along with a Hue smart lights bridge, which is paired with a couple of Arduinos and 2.4GHz radio modules that handle the wireless communication between the two devices.

You can see more about this build on their blog here and more background on their lighting system here.

Motorize any camera slider with Arduino!

Using a stepper motor, an Arduino Micro, and various mechanical bits, “Max Maker” shows us how to motorize a camera slider.

According to his video below, “There are many cheap sliders available, but none that are motorized.” On the other hand, in this project he demonstrates that with a little bit of work, and an excellent attention to detail, you can make the conversion yourself.

Even if you’re not interested in this kind of application, the video reveals some interesting tricks, such as transferring a hole pattern using tape at 0:30, and using nail polish/recessed lettering to label switches at 3:55. The slider can be set to slide down the rail between 10 seconds and 8 hours depending on your video or photography needs. Video results, seen around 5:00 in, are really amazing!

You can find more details on how to create a slider like this on the project’s Instructables page.

An experimental game with a conductive rubber band controller

RubberArms is an experimental rubber band game, created by Robin Baumgarten at the Global Game Jam 2017 in Yverdon-les-Bains, Switzerland.

The controller uses a conductive rubber cord from Adafruit that changes resistance as it’s stretched. This resistance is measured by an Arduino Micro/Leonardo (or a Teensy 3.2), which acts as a USB joystick sending signals to Unity3D. (The game is coded in Unity3D using Spring Joints and Line Renderers.)

At this point, the game is a simple prototype where you control the distance of two characters whose arms stretch whenever you stretch the rubber band, throwing little ‘Bleps’ around. You can read more about RubberArms on Baumgarten’s page, as well as his earlier project “Line Wobbler” here.

This 3D-printed bionic hand can replace or support a limb

3D-printed appendages are, as one might suspect, generally meant for those that are missing a limb. Moreover, there are many other people that might retain partial functionality of a hand, but could still use assistance.

Youbionic’s beautifully 3D-printed, myoelectric prosthesis is envisioned for either application, capable of being controlled by muscle contraction as if it were a real body part.

As seen in the video below, the Youbionic hand can manipulate many different items, including a small box, a water bottle, and a set of keys. Functionality aside, the movement is extremely fluid and the smooth black finish really makes it look great.

The device is currently equipped with an Arduino Micro, servos, various sensors, a battery pack, and a few switches. Even the breadboard appears to be very neat, though one would suspect the final version will use some sort of PCB.

You can learn more and order yours on Youbionic’s website.

Smart sock augments existing prostheses’ abilities

Developed by researchers at the University of Applied Sciences in Linz, the proCover is a sensor-enabled smart sock that adds sensations to current prosthetic limbs.

Although work on more advanced prostheses continues, commercially available limbs still lack tactile feedback. The proCover addresses this not by modifying or replacing the prosthesis, but by using a sock with piezoresistive force sensors embedded in it. This allows for the user to tell where on a foot it’s being touched, as well as the pressure applied, and it can be set up to suit a user’s needs.

Feedback is provided by vibrating rings that can also be placed on a user’s body where convenient. A version that detects how far a prosthetic knee is bent has also been tested.

The design and construction of prostheses that can emulate a natural sense of touch is of growing research interest. Over the last few decades, a number of solutions have been developed for the detection of pressure, slip, heat and texture… However, many of the exciting innovations in this field will likely remain out of reach for most people due to a multitude of factors pertaining to cost, accessibility, health status, and personal attitudes towards elective surgery. We introduce proCover, a low-cost sensing wearable in the form of a textile sock that can be applied retroactively to lower-limb prosthetics to make sensing capabilities more broadly accessible to those who rely on these assistive devices.

You can find more information in the project’s 12-page paper or in this TechCrunch article for a slightly shorter summary.

(Photos: Media Interaction Lab)

Smart sock augments existing prostheses’ abilities

Developed by researchers at the University of Applied Sciences in Linz, the proCover is a sensor-enabled smart sock that adds sensations to current prosthetic limbs.

Although work on more advanced prostheses continues, commercially available limbs still lack tactile feedback. The proCover addresses this not by modifying or replacing the prosthesis, but by using a sock with piezoresistive force sensors embedded in it. This allows for the user to tell where on a foot it’s being touched, as well as the pressure applied, and it can be set up to suit a user’s needs.

Feedback is provided by vibrating rings that can also be placed on a user’s body where convenient. A version that detects how far a prosthetic knee is bent has also been tested.

The design and construction of prostheses that can emulate a natural sense of touch is of growing research interest. Over the last few decades, a number of solutions have been developed for the detection of pressure, slip, heat and texture… However, many of the exciting innovations in this field will likely remain out of reach for most people due to a multitude of factors pertaining to cost, accessibility, health status, and personal attitudes towards elective surgery. We introduce proCover, a low-cost sensing wearable in the form of a textile sock that can be applied retroactively to lower-limb prosthetics to make sensing capabilities more broadly accessible to those who rely on these assistive devices.

You can find more information in the project’s 12-page paper or in this TechCrunch article for a slightly shorter summary.

(Photos: Media Interaction Lab)

LipSync is an Arduino-based assistive device for smartphone use

With a mouth-operated joystick and “sip and puff” controls, the LipSync aims to make smartphones more accessible for everyone.

For the huge number of people that use them, smartphones have certainly made their lives easier. Unfortunately, these amazing gadgets are difficult to use for those with limited or nonexistent use of their arms and hands. The LipSync attempts to address this issue with a device that can be made in just over a weekend’s worth of work. It uses an Arduino Micro along with a Bluetooth module for communication, and allows someone to interface with the phone using its tiny joystick, as well as the user’s controlled breath.

Smartphones and other similar mobile devices have become a staple piece of technology in this day and age. For people in wheelchairs whom experience difficulties with gross or fine upper body motor control, the usage of mobile devices can be very challenging. The LipSync is an assistive technology device which is being developed to allow quadriplegics the ability to use touchscreen mobile devices by manipulation a mouth-operated joystick with integrated sip and puff controls.

You can find more information on this project, including the files needed to build one, on its Hackaday.io page.

Drive an RC car with first-person view

If you’ve ever dreamt of sitting in the driver’s seat of an RC car, you’ll love this recent project from Paul Yan. The designer has hacked together a first-person driving experience using a natural steering wheel that lets you feel like you’re playing a race kart game in real life.

As he describes in the video below, Yan used an old PS2 wheel controller, two Arduinos, a mini FPV camera, and a headset to act as a standalone monitor. The RC car–which is equipped with a Micro–interfaces with the wheel using an Uno and a PS2 Shield. Both Arduinos communicate via a pair of NRF24L01 modules.

I got the idea for this project while watching my three-year-old son play with his radio controlled toy cars. These all use the conventional two joysticks. The left is isolated to vertical movement to control the motor and the right stick is isolated to horizontal movement to control the turning direction of the front wheels.

Sorting cucumbers using AI, Raspberry Pi + Arduino

When it comes to farming veggies like cucumbers, the sorting process can often be just as hard and tricky as actually growing them. That’s why Makoto Koike is using Google’s TensorFlow machine learning technology to categorize the cucumbers on his family’s farm by size, shape and color, enabling them to focus on more important and less tedious work.

A camera-equipped Raspberry Pi 3 is used to take images of the cucumbers and send them to a small-scale TensorFlow neural network. The pictures are then forwarded to a larger network running on a Linux server to perform a more detailed classification. From there, the commands are fed to an Arduino Micro that controls a conveyor belt system that handles the actual sorting, dropping them into their respective container.

You can read all about the Google AI project here, as well as see it in action below!