Posts with «art» label

The Battle Between Robot Harmonica And Machine Finger Rages On

When asking the question “Do humans dream of machines?”, it’s natural to think of the feverish excitement ahead of an iPhone or Playstation launch, followed by lines around the block of enthusiastic campers, eager to get their hands on the latest hardware as soon as is humanly possible. However, it’s also the title of an art piece by [Jonghong Park], and is deserving of further contemplation. (Video after the break.)

The art piece consists of a series of eight tiny harmonicas, which are in turn, played by eight fans, which appear to have been cribbed from a low-power graphics card design. Each harmonica in turn has a microphone fitted, which, when it picks up a loud enough signal, causes an Arduino Nano to actuate a mechanical finger which slows the fan down until the noise stops. It’s the mechanical equivalent of a stern look from a parent to a noisy child. Then, the cycle begins again.

The build is very much of the type we see in the art world – put together as simply as possible, with eight Arduinos running the eight harmonicas, whereas an engineering approach may focus more on efficiency and cost. Between the squeaks from the toy harmonicas and the noise from the servos entrusted to quiet them, the machine makes quite the mechanical racket. [Jonghong] indicates that the piece speaks to the interaction of machine (robot harmonica) and humanity (the finger which quells the noise).

It’s a tidily executed build which would be at home in any modern art gallery. It recalls memories of another such installation, which combines fans and lasers into a musical machine.

 

Hack a Day 13 Dec 19:30

Artistic Images Made With Water Lens

It’s said that beauty and art can be found anywhere, as long as you look for it. The latest art project from [dmitry] both looks in unassuming places for that beauty, and projects what it sees for everyone to view. Like most of his projects, it’s able to produce its artwork in a very unconventional way. This particular project uses water as a lens, and by heating and cooling the water it produces a changing image.

The art installation uses a Peltier cooler to periodically freeze the water that’s being used as a lens. When light is projected through the frozen water onto a screen, the heat from the light melts the water and changes the projected image. The machine uses an Arduino and a Raspberry Pi in order to control the Peliter cooler and move the lens on top of the cooler to be frozen. Once frozen, it’s moved again into the path of the light in order to show an image through the lens.

[dmitry] intended the project to be a take on the cyclical nature of a substance from one state to another, and this is a very creative and interesting way of going about it. Of course, [dmitry]’s work always exhibits the same high build quality and interesting perspective, like his recent project which created music from the core samples of the deepest hole ever drilled.

Hack a Day 08 Dec 03:00
arduino  art  cooler  ice  lens  peltier  projection  raspberry pi  water  

Kinetic Sculpture Achieves Balance Through Machine Learning

We all know how important it is to achieve balance in life, or at least so the self-help industry tells us. How exactly to achieve balance is generally left as an exercise to the individual, however, with varying results. But what about our machines? Will there come a day when artificial intelligences and their robotic bodies become so stressed that they too will search for an elusive and ill-defined sense of balance?

We kid, but only a little; who knows what the future field of machine psychology will discover? Until then, this kinetic sculpture that achieves literal balance might hold lessons for human and machine alike. Dubbed In Medio Stat Virtus, or “In the middle stands virtue,” [Astrid Kraniger]’s kinetic sculpture explores how a simple system can find a stable equilibrium with machine learning. The task seems easy: keep a ball centered on a track suspended by two cables. The length of the cables is varied by stepper motors, while the position of the ball is detected by the difference in weight between the two cables using load cells scavenged from luggage scales. The motors raise and lower each side to even out the forces on each, eventually achieving balance.

The twist here is that rather than a simple PID loop or another control algorithm, [Astrid] chose to apply machine learning to the problem using the Q-Behave library. The system detects when the difference between the two weights is decreasing and “rewards” the algorithm so that it learns what is required of it. The result is a system that gently settles into equilibrium. Check out the video below; it’s strangely soothing.

We’ve seen self-balancing systems before, from ball-balancing Stewart platforms to Segway-like two-wheel balancers. One wonders if machine learning could be applied to these systems as well.

Salvaged Arduino powers animated House Party

What can you do with items that are destined for the dump? As seen here, if you’re Neil Mendoza, you transform furniture, an old TV, art, and even an Arduino Zero that somehow ended up in the trash into a musical installation.

His resulting “House Party” features decorations and control components that according to the project’s write-up are entirely salvaged. A MIDI interface, software written in openFrameworks, and a JSON file are used to coordinate sound and movements, which include spinning picture frames and flowers, tapping shoes, and a television that loops through a rather dreary weather report snippet. 

House Party is a musical installation that explores prized possessions in their native habitat. All the materials used to create this artwork, from the furniture to the computers, were scavenged from the discarded trash. The music is a mix of mechanical and synthesized sounds. The piece was created while an artist in residence at Recology SF.

The actuators in the installation are controlled by an Arduino Zero (also found in the trash) and each screen is connected to a computer running custom software written in openFrameworks (OF). Composition was done in Logic where a MIDI environment was set up to send MIDI data to the Arduino and an OF control program. The control program then sent the data to the other computers over ethernet as OSC. For the final installation, the control program read the data from a JSON file, triggered the screens and Arduino and played the synthesized parts of the music.

Be sure to see all the zany action in the video below!

Redeem Your Irresponsible 90s Self

If you were a youth in the 90s, odds are good that you were a part of the virtual pet fad and had your very own beeping Tamagotchi to take care of, much to the chagrin of your parents. Without the appropriate amout of attention each day, the pets could become sick or die, and the only way to prevent this was to sneak the toy into class and hope it didn’t make too much noise. A more responsible solution to this problem would have been to build something to take care of your virtual pet for you.

An art installation in Moscow is using an Arduino to take care of five Tamagotchis simultaneously in a virtal farm of sorts. The system is directly wired to all five toys to simulate button presses, and behaves ideally to make sure all the digital animals are properly cared for. Although no source code is provided, it seems to have some sort of machine learning capability in order to best care for all five pets at the same time. The system also prints out the statuses on a thermal printer, so you can check up on the history of all of the animals.

The popularity of these toys leads to a lot of in-depth investigation of what really goes on inside them, and a lot of other modifications to the original units and to the software. You can get a complete ROM dump of one, build a giant one, or even take care of an infinite number of them. Who would have thought a passing fad would have so much hackability?

Hack a Day 19 Oct 06:00

Pay for art with your mugshot

As reported here, digital artist Matthias Dörfelt has created an art vending machine in an attempt to increase awareness around blockchain possibilities, as well as how we handle our personal information.

Face Trade, now on display at Art Center Nabi in Seoul, takes the form of a large vaguely face shaped box. When it detects a human in front of it, the installation invites the participant to swap his or her face for art, confirmed using a large yellow button that connects to the system’s computer via an Arduino.

Once confirmed, Face Trade snaps the person’s picture and uploads it to a blockchain in exchange for a computer generated facial image. The resulting art’s conflicted expression is meant to signify the good and bad possibilities that can come out of using this technology. For their trouble, participants also get a receipt showing their captured headshot that now appears along with each transaction on itradedmyface.com.

Face Trade consists of a camera flash, webcam, receipt printer, inkjet printer, computer, speakers, LCD screen, button and an Arduino (to control the button, LCD screen and camera flash).

The main application that ties everything together is written in Python. It uses OpenCV to do basic face tracking and take the images. All the Ethereum related things were done using web3.py which is the official python version of web3 to interact with the Ethereum blockchain. The receipt printer, inkjet and Arduino are controlled via Python, too. The process is comprised of taking a picture, uploading it to the blockchain, passing the resulting transaction hash to the face drawing generator that uses it to seed the random numbers (so that each face drawing is uniquely tied to the transaction that it belongs to), printing the resulting drawing and finally printing the receipt.

Shy robotic sculpture imitates nature

In nature, animals often are sensitive to the outside environment, retreating into a hole, shell, or other protective structure upon sensing sudden movements. If you were to envision this kind of behavior in robot form, you might come up with something like “The Shy Machine” from Daric Gill Studios.

When it detects motion via a PIR sensor, the shell-like robot takes a reading of the ambient sound level using an internal microphone. If things are sufficiently quiet, it opens up using a stepper motor and lead screw, revealing a rainbow of colors provided by an array of RGB LEDs inside.

Its construction and a demo video are shown below, and you can see more about how this Arduino-powered robot was built and the results on Gill’s website.

RGB Disk Goes Interactive with Bluetooth; Shows Impressive Plastic Work

[smash_hand] had a clear goal: a big, featureless, white plastic disk with RGB LEDs concealed around its edge. So what is it? A big ornament that could glow any color or trippy mixture of colors one desires. It’s an object whose sole purpose is to be a frame for soft, glowing light patterns to admire. The disk can be controlled with a simple smartphone app that communicates over Bluetooth, allowing anyone (or in theory anything) to play with the display.

The disk is made from 1/4″ clear plastic, which [smash_hand] describes as plexiglass, but might be acrylic or polycarbonate. [smash_hands] describes some trial and error in the process of cutting the circle; it was saw-cut with some 3-in-1 oil as cutting fluid first, then the final shape cut with a bandsaw.

The saw left the edge very rough, so it was polished with glass polishing compound. This restores the optical properties required for the edge-lighting technique. The back of the disc was sanded then painted white, and the RGB LEDs spaced evenly around the edge, pointing inwards.

The physical build is almost always the difficult part in a project like this — achieving good diffusion of LEDs is a topic we talk about often. [smash_hands] did an impressive job and there are never any “hot spots” where an LED sticks out to your eye. With this taken care of, the electronics came together with much less effort. An Arduino with an HC-05 Bluetooth adapter took care of driving the LEDs and wireless communications, respectively. A wooden frame later, and the whole thing is ready to go.

[smash_hands] provides details like a wiring diagram as well as the smartphone app for anyone who is interested. There’s the Arduino program as well, but interestingly it’s only available in assembly or as a raw .hex file. A video of the disk in action is embedded below.

Making LED lighting interactive comes in many different shapes and forms, and as the disk above shows, shifting color patterns can be pleasantly relaxing.

Follow the Bouncing Ball of Entropy

When [::vtol::] wants to generate random numbers he doesn’t simply type rand() into his Arduino IDE, no, he builds a piece of art. It all starts with a knob, presumably connected to a potentiometer, which sets a frequency. An Arduino UNO takes the reading and generates a tone for an upward-facing speaker. A tiny ball bounces on that speaker where it occasionally collides with a piezoelectric element. The intervals between collisions become our sufficiently random number.

The generated number travels up the Rube Goldberg-esque machine to an LCD mounted at the top where a word, corresponding to our generated number, is displayed. As long as the button is held, a tone will continue to sound and words will be generated so poetry pours forth.

If this take on beat poetry doesn’t suit you, the construction of the Ball-O-Bol has an aesthetic quality that’s eye-catching, whereas projects like his Tape-Head Robot That Listens to the Floor and 8-Bit Digital Photo Gun showed the electronic guts front and center with their own appeal.


Filed under: Arduino Hacks

GuitarBot Brings Together Art and Engineering

Not only does the GuitarBot project show off some great design, but the care given to the documentation and directions is wonderful to see. The GuitarBot is an initiative by three University of Delaware professors, [Dustyn Roberts], [Troy Richards], and [Ashley Pigford] to introduce their students to ‘Artgineering’, a beautiful portmanteau of ‘art’ and ‘engineering’.

The GuitarBot It is designed and documented in a way that the three major elements are compartmentalized: the strummer, the brains, and the chord mechanism are all independent modules wrapped up in a single device. Anyone is, of course, free to build the whole thing, but a lot of work has been done to ease the collaboration of smaller, team-based groups that can work on and bring together individual elements.

Some aspects of the GuitarBot are still works in progress, such as the solenoid-activated chord assembly. But everything else is ready to go with Bills of Materials and build directions. An early video of a strumming test proof of concept used on a ukelele is embedded below.

GuitarBot would fit right in to a band where only the instruments operate unplugged. Speaking of robot bands, don’t forget the LEGO-enabled Toa Mata, or the fully robotic group Compressorhead.


Filed under: musical hacks, robots hacks