As first reported by Technabob, almost two years after he made a rather stunning Saturn V lamp, SimonRob decided to create something a bit different in the form of his Saturn V planter.
The device is based on the lower section portion of a 3D-printable rocket model, which is modified and sealed to hold dirt and water. A succulent now pops out where the rest of the rocket should be, and when combined with a 30mm thick wood plank, gives it a very unique and polished look.
A switch on the front lights up the printed flame assemblies emanating from the engines, using a trio of SMD LEDs on each exhaust. These LEDs are controlled by an Arduino Nano nestled inside the wooden base to produce random lighting effects when an activation button is pressed.
When you need to grab someone’s attention at an event, an interactive screen is a good idea. MakerMan, however, went several steps beyond this, creating an installation with a bank of static screens that depict the Moscow skyline. In front of this, a single touch-enabled display moves back and forth automatically to present information on various points of interest.
Sliding action is handled by a large stepper motor, which pulls the screen along on a carriage assembly. The motor, in turn, is controlled via an Arduino Uno and a stepper driver. All of these electronics are hidden behind a nicely painted wooden facade, letting the technology driving it fade elegantly into the background.
After studying the way a worm wiggles, Nicholas Lauer decided to create his own soft robotic version. What he came up with uses an Arduino Uno for control, inflating six 3D-printed segments sequentially to order to generate peristaltic motion for forward movement.
The robotic worm uses a 12V mini diaphragm pump to provide inflation air, while a series of transistors and solenoid valves directly regulate the airflow into the chambers.
The build looks pretty wild in the video below, and per Lauer’s write-up, you’re encouraged to experiment to see what kind of timing produces the most expedient motion. Code, STLs, and a detailed BOM are available on GitHub.
While taking photos today is normally a digital affair, there is a wealth of visual information stored on film negatives. Digitization is possible, but it tends to be rather time-intensive, so photographer/hacker Seckin Sinan Isik decided to automate the process.
His setup uses a film carrier augmented with a stepper motor and belt drive to advance the 35mm film under a tripod-mounted digital camera. This is controlled by an Arduino Nano, with the camera’s view shown via a video capture device on a nearby computer.
In one mode, the user can adjust the film position semi-manually using pushbuttons, then scan the negative. The whole process can also be automated, with a Python computer vision routine.
Learning a new language is always a challenge, but can also be extremely rewarding. To help with this task — specifically learning Greek — Angeliki Beyko created an arcade-style review machine. Her device cleverly uses a sheet of pegboard to hold the electronics, including seven LCD screens to indicate category, level, and the actual Greek word being reviewed.
Possible answers are shown as pictures on four TFT displays, driven by four separate Arduino Uno boards that pull up pictures stored on SD cards. An Arduino Mega provides overall control for the interactive panel.
While ambitious, and a definite learning process, there’s currently some memory limitations and an issue with the screens not working when actually mounted. So as of now, it’s something of a work-in-progress, but Beyko is planning to complete the build with version two. More details are available in the project write-up and code can be found on GitHub.
Humans are generally quite bad at coming up with random patterns, so when Jeremy Cook wanted to make a sound diffuser with angled blocks of wood, he created a “pseudorandomness console” using an Arduino Uno and an LCD shield.
This helped him with the placement of its 216 angled segments, which are colored in one of four ways, and can face up, down, left, and right to theoretically scatter sound in every direction.
Code for this unique randomization is available on GitHub, with a quick explanation in the video above. You can see the final assembly at around the 4:38 mark, showing a process of applying glue, pressing a button to generate a value, and then placing triangles accordingly.