Below is a detailed tutorial on how to create a data-moshing program and run it from a raspberry pi. Every time the program runs, the data-moshed photo is uploaded to the portfolio google site linked here.
What exactly is this project anyway?
- Essentially this project runs a few different functions I have written to combine the data from photos I took over the summer. Most of the functions are pretty simple and I am always looking to add more to them. Once the functions have been applied to the image data, I am left with a really cool, unique photo. Spontaneity is built into the core of each function so I get a completely different result every time the data bending script is run. One of the functions I have written pastes data from the MediaWiki API over the image. (By the way, data bending and data moshing are just a short term for editing photos in unexpected ways)
Motivation
- I have always been fascinated with the concept of manipulating data in visual ways. I like being able to see what is happening and turn it in to an art form. This project is actually a school data visualization project, and I decided that it would be really interesting to use the photos I had taken throughout the whole of quarantine in my project. I picked up photography during the summer of Covid-19 since I can do it without interacting with many people at all, and this data visualization project is the perfect place for me to use all the photos I have taken to make something new.
I used five different libraries to make this project possible: requests, json, random, selenium, and PIL. The selenium installation process on the raspberry pi is different than on a Mac or Windows machine. I have more details on how to properly install selenium to the raspberry pi under "Migrating to the Raspberry Pi"; everything in this section about selenium will be about installing it to a mac or windows machine. I used the following import statements to import the parts of the Libraries I would need.
import requests
import json
import random
import selenium
from PIL import Image, ImageDraw
Importing a library at the top of your code and actually having it installed, however, are two separate things. Before I could use these libraries I had to install them, except for random since it comes default with Python 3. The two ways I know of installing Libraries are through PyCharm and through PIP. I actually had to use both of these methods during this project when I learned that the version of PyCharm loaded on my spare computer does not have the feature that auto-installs libraries.
PyCharm
- To install the necessary Libraries using PyCharm, type the import statements pictured above and right-click each one. The IDE should tell you something along the lines that it does not recognize the import statement. There should be an option to install the library you are attempting to import. If the such an option does not exist, you will need to use PIP to install your libraries.
- Selenium requires a ChromeDriver to be installed in addition to the library. To the best of my knowledge, PyCharm is unable to install this. Follow the links at the bottom of the page to manually install the ChromeDriver
PIP
- Installing the necessary Libraries using PIP will be a bit more involved but still very easy. I used the commands below to install PIL, requests, and Selenium using pip when I had to migrate to the raspberry pi later on.
- Here are the commands for installing PIL:
$ python3 -m pip install --upgrade pip
$ python3 -m pip install --upgrade Pillow
- Here are the commands for installing requests (use one of them):
$ python -m pip install requests
$ pip3 install requests
- Here is the command for installing selenium:
$ pip install selenium
- selenium also requires a ChromeDriver to be installed follow the links at the bottom of this section for a guide on how to properly install the ChromeDriver. I have also linked the ChromeDriver download site below.
- Since I am not an expert on this and it can vary depending on the machine you are using and what you have installed on it, I suggest reading more from the links below on how to install pip if you do not already have it and how to install the necessary libraries for this project with it.
- PIP Installation guide: https://pypi.org/project/pip/
- PIL Installation guide: https://pillow.readthedocs.io/en/stable/installation.html
- Requests Installation guide: https://pypi.org/project/requests/
- Selenium Installation guide: https://selenium-python.readthedocs.io/installation.html
- ChromeDriver Installation guide: https://chromedriver.chromium.org/getting-started
- ChromeDriver Installation link: https://sites.google.com/a/chromium.org/chromedriver/downloads
I used the MediaWiki Action API to grab the introductions of specified Wikipedia pages so that I could overlay the text on one of my data moshed images. The Wikipedia API was actually pretty easy to use. The biggest challenge was finding the right parameters to pass in my request. The MediaWiki documentation was sparse in areas and a bit over my head, but I eventually learned that the extracts
property was what I needed. I was only able to learn about this thanks to the Stack Overflow page linked below. I also have the MediaWiki API documentation page linked below since it gave me the basics to understanding the API. I suggest reading through the documentation to understand how to make a request to the API and to learn more about the parameters being passed.
An example of a finalized API request:
"https://en.wikipedia.org/w/api.php?"
"format=json&"
"action=query&"
"prop=extracts&"
"exintro&explaintext&"
"redirects=1&"
"titles=Stack%20Overflow"
The above request's response:
{
"batchcomplete":"",
"query":{
"pages":{
"21721040":{
"pageid":21721040,
"ns":0,
"title":"Stack Overflow",
"extract":"Stack Overflow is a question and answer site ... open to being contacted."
}
}
}
}
(I have shortened the above extract to make the page legible. There is no ellipsis in the actual API response.)
I had my code request the introduction page from a random page on my list of select Wikipedia pages using the code below. The code uses a basic request and concatenates a random string from the list titles to the end.
def formulate_query() -> str:
query = ("https://en.wikipedia.org/w/api.php?"
"format=json&"
"action=query&"
"prop=extracts&"
"exintro&explaintext&"
"redirects=1&"
"titles=")
titles = ["Stack%20Overflow",
"Industrialisation",
"Zen%20and%20the%20Art%20of%20Motorcycle%20Maintenance",
"Computer%20science",
"Smog",
"Machine",
"Sentience"]
return query + titles[random.randrange(0, len(titles))]
- https://stackoverflow.com/questions/8555320/is-there-a-wikipedia-api-just-for-retrieve-content-summary.
- https://www.mediawiki.org/wiki/API:Main_page
I utilized the Image and the ImageDraw Modules from PIL to respectively edit the images and overlay the text from Wikipedia. This section explains what the image editing functions do and how they do it. If you are interested in learning more about how I used PIL and how the functions I have written work, then continue reading. If not, skip to "Migrating to the Raspberry Pi".
ImageModule
- The Image module allowed me to access, edit, and save the image, and the program mostly works with the List of pixels from each image. The pixels are formatted in Tuples of three inside of a one-dimensional list. An example of a white two by two pixel image is below
data = [(255, 255, 255), (255, 255, 255), (255, 255, 255), (255, 255, 255)]
- The functions that interacted directly or near directly with the Image module are such.
def initializeImg(path: str) -> object:
def resizeImg(img: Image, size: str) -> Image:
def tolist(img: Image) -> list:
def formulate_path() -> str:
def saveImg(img: Image) -> None:
def bend(img1: Image, img2: Image, data1: list, data2: list) -> Image:
DataMoshingFunctions
- Bend - This function is the main data bending function. Its main purpose is to call a randomized amount of data moshing functions in a randomized order. The function uses the dictionary params and the list of functions f to do this. Since each function has a different set of parameters, it is easiest to just pass all of the parameters in a dictionary when calling the functions randomly in the for loop. Once the loop is complete and
params['data1']
has been modified by the functions, theputdata
function is called and the one-dimensional array passed asdata
overwrites the pixels of the original image.
def bend(img1: Image, img2: Image, data1: list, data2: list) -> Image:
params = {'img1': img1, 'img2': img2, 'data1': data1, 'data2': data2}
f = [drag, corrupt, contrast, contrast_blue, contrast_green, contrast_red]
for i in range(random.randint(3, 8)):
params['data1'] = f[random.randint(0, 5)](params)
img1.putdata(data=params['data1'])
return img1
- Spread - This function scatters the pixels by within a randomized radius. It essentially blurs the image. I do not like how this makes the image look so I decided to remove
spread
from the list of functionsf
. But, since I already wrote the function and beauty is in the eye of the beholder, I left this function in the code on this project. Feel free to add it tof
and experiment!
def spread(params: dict) -> Image:
img = params['img1']
return img.effect_spread(random.randint(5, 25))
- Weave - This function basket-weaves the two images together. The sub-function
find_factors
finds all the factors of the image width and height excluding 1 and the image width or height since we don't want the first cell of the basket-weave to be the width of the image. The variablewidth_divisor
is the number of cells per row, and the variablecube_width
is the width in pixels of each cell. Consequently, height_divisor is the number of cells per column andcube_height
is the height in pixels of each cell. The purpose of the outermost loop,for d in range(height_divisor):
, is to loop through the number of cubes or cells that fit in the image's height. At the end of every iteration, the loop inverses the value of n ensuring that the cubes of img1 data and img2 data will alternate. Since most of the factors of the images width and height are even numbers, the cells would not alternate on the vertical axis (I will include an image of this below the code snippet). The next loop ,for h in range(cube_height):
, loops through the cube_height building each cube/cell row by row. The following loop,for w in range(width_divisor):
, loops through the number of cubes that fit in the img width. At the end of every iteration, this loop changes the state of n so that the inner-most loop will only write img2 data over the img1 data on every other iteration. Finally, the inner-most loop,for i in range(start, stop):
, loops through the width of a cell writing pixel data from img2 to img1 for every other cube_width (i.e. when n == True).
def weave(params: dict) -> list:
data1 = params['data1']
data2 = params['data2']
def find_factors(num: int) -> list:
factors = []
for f in range(2, num):
if num % f == 0:
factors.append(f)
return factors
width_factors = find_factors(params['img1'].size[0])
width_divisor = width_factors[random.randint(0, len(width_factors) - 1)]
cube_width = params['img1'].size[0] // width_divisor
height_factors = find_factors(params['img1'].size[1])
height_divisor = height_factors[random.randint(0, len(height_factors) - 1)]
cube_height = params['img1'].size[1] // height_divisor
start = 0
stop = cube_width
n = True
for d in range(height_divisor):
for h in range(cube_height):
for w in range(width_divisor):
for i in range(start, stop):
if n:
data1[i] = data2[i]
else:
break
n = not n
start = start + cube_width
stop = stop + cube_width
n = not n
return data1
- Drag - This function takes a random row of pixels and copies it to the adjacent rows below a randomized amount of times. It looks like the row of pixels has been dragged down hence the function name. In this function,
c
determines the height of the dragged portion in pixels,a
determines the starting point of the dragged portion starting anywhere from the upper left-hand corner toc
pixels from the bottom most row,b
determines the last pixel plus one in a row so that the data can be copied to lower rows.
def drag(params: dict) -> list:
img = params['img1']
data = params['data1']
width, height = img.size
c = random.randint(1, height / 2)
a = random.randrange(0, len(data) - (width * c), width)
b = a + width
for i in range(1, c):
for j in range(a, b):
data[j + width * i] = data[j]
return data
- Corrupt - This function copies sections of the second photo onto the first photo. In this function,
a
determines the beginning of the copied data andb
determines the end of the copied data.b
will always be some number larger thana
even if it is only by one. This is ensured by the parameters of therandom.randint()
function. I was fairly sure that my code would not cause anIndexError
by trying to access an element indata1
that does not exist, but I added the try and except block just in case and had it return data1 with all of the changes made to it leading up to the error. The for loop sets the pixel values at indexi
from the first image to the pixel value at indexi
from the second image.
def corrupt(params: dict) -> list:
data1 = params['data1']
data2 = params['data2']
a = random.randint(0, len(data1) - 1)
b = random.randint(a, len(data1))
for i in range(a, b):
try:
data1[i] = data2[i]
except IndexError:
print("IndexError occurred " + str(IndexError) + ". Returning data1 as is.")
return data1
return data1
- Contrast - This function contrasts each pixel value in
data1
by subtracting it from 255. If a color was maximized at 255, it becomes minimized at 255-255 or 0. This pure red (255, 0, 0) becomes pure yellow (0, 255, 255). Since I have four of these functions where three of them only contrast one color, I will only include the explanation and code of this function.
def contrast(params: dict) -> list:
data = params['data1']
for i in range(len(data)):
data[i] = (255 - data[i][0], 255 - data[i][1], 255 - data[i][2])
return data
ImageDrawModule
- The ImageDraw module allowed me to overlay text on my images. I had to create a
Draw
object to do this. - The functions that interacted directly or almost directly with this module were such
def initializeDraw(img: Image) -> ImageDraw:
def convert_to_drawable(response: str) -> str:
def drawText(draw: ImageDraw) -> None:
TextOverlayFunctions
- initializeDraw - This function creates and returns a
Draw
object that is connected to one of theImage
objects passed as a parameter.
def initializeDraw(img: Image) -> ImageDraw:
return ImageDraw.Draw(img)
- convert_to_drawable - This function interacts more with the MediWiki API than the ImageDraw module, but it is essential in making the text fit within the image. This function adds line breaks
\n
at every period so that most of the text fits on the image. I thought about making a proper solution to the text trailing off the image, but I actually like that the text can disappear off the screen on one line and be really short on another. I think it adds to the chaos of the images.
def convert_to_drawable(response: str) -> str:
split_response = response.split('.')
text = ""
for i in split_response:
text = text + "\n" + i
return text
- draw_text - This function draws or overlays the text onto the image that the Draw object was initialized with. The
(0, 0)
parameter determines at what pixel the text begins, theconvert_to_drawable(request())
is actually the text that is overlaid on the image, and thefill
parameter determines the color and opacity of the text.
def drawText(draw: ImageDraw) -> None:
draw.multiline_text((0, 0), convert_to_drawable(request()), fill=(255, 25 5, 255, 255))
Migrating to the Raspberry PiOnce I had finished the program, I moved it to my raspberry pi 3 B. The easiest way to do this, that I have found, was simply to transfer the program via flash drive. Here is a photo of the program on my raspberry pi:
Once again, I had to install the libraries that the scripts use. This time I had to use PIP. The following commands are what worked on my raspberry pi:
$ python3 -m pip install --upgrade pip
$ python3 -m pip install --upgrade Pillow
Follow the following link for the PIL Installation guide using pip: https://pillow.readthedocs.io/en/stable/installation.html
CUPS Printing ScriptThis script is entirely optional and can be swapped out or used in unison with the Google Site portfolio scripts. Its purpose is to print the the result of the data moshing script. This project functionality only works on the Raspberry Pi.
Installing CUPS on the Raspberry Pi:
To install the library, run the following command in the terminal window.
$ sudo apt-get install python-cups
Connecting The Printer:
Before being able to use the CUPS functions getPrinters()
and printFile()
, the printer has to be connected in some way to the Raspberry Pi. I was able to connect my printer through the wifi shown in the photo below.
Once the printer is connected, the commented section of the script can be run to determine the name of the printer you want to use. Once you know the printer's name, you can re-comment that section of the code and enter the name into the printFile()
function. Then you are good to go. I have included some photos below to document my printing process as well as the sources I have used. I can imagine that connecting a printer to the Raspberry Pi is a finicky process, so I suggest following these sources to read more about how CUPS works and how to get started using it.
Sources:
A big thanks to the sources bellow, I would not have been able to figure this out on my own.
- https://pypi.org/project/pycups/
- https://www.techradar.com/how-to/computing/how-to-turn-the-raspberry-pi-into-a-wireless-printer-server-1312717/2
- https://www.raspberrypi.org/blog/printing-at-home-from-your-raspberry-pi/
The idea is that the result of the data moshing script will be uploaded to a Google Site using Selenium.
The raspberry pi must first be fully up to date which we can do with the following terminal command:
$ sudo apt-get update
Then you will need to install Selenium.
$ sudo pip3 install selenium
Chromium is not officially supported on Raspbian, but there is a work around through https://launchpad.net/ubuntu/xenial/armhf/chromium-chromedriver. It is important that you install the version of chromium that matches your installed chromium-browser version which can be checked with the following command.
$ chromium-browser --version
The version I had installed was 84.0.4147.141 so I went ahead and downloaded the chrome-driver version that matched (or most nearly matched since I downloaded version 84.0.4147.105). Once you have the chromium chrome-driver downloaded, you will need to move the unzipped file into /usr/bin/
. After this Selenium is good to go.
You will need to create a new google site before being able to use Selenium to automate the uploading of photos to the portfolio site and the publishing of that site. After you have done this, you are ready to use the Upload Image
and Publish Site
scripts.
Upload Image Script:
The upload image script is run with os.system()
at the very end of the Data Mosh script. This script is able to log in to the edit page of the Google Site and upload the result of the Data Mosh script. I ran into an error where the file manager window stayed open after the image was uploaded to the Google Site halting all further actions of the selenium web-driver. A work-around for this that I have been using is to quit the web-driver with driver.quit()
and use a second script for the following actions. This second script is the Publish Site script.
Publish Site Script:
The Publish Site script is run the same way as the Upload Image script at the very end of the Upload Image script. All this script does is log into the Google Site edit page and click the 'Publish' button followed by the second 'Publish' button. Once the site has been published, the web-driver is quit.
Side Note:
It is important to use the link directly to the edit page of your portfolio Google Site. Both the Upload Image and Publish Site scripts assume this is the initial link being opened by the web-driver with driver.get()
. The scripts will not work unless this link is used.
A big thanks to the following sources:
- https://www.youtube.com/watch?time_continue=446&v=6PBmkNkfglk&feature=emb_title
- https://www.youtube.com/watch?v=4TuawlLrWaY
- https://launchpad.net/ubuntu/xenial/armhf/chromium-chromedriver
- https://github.com/ccozkan/selenium-instagram-uploader
Now enough with the technical talk; here is a progression of the programs output as well as my favorite one. When Testing I worked with smaller image sizes so I could run the program much more quickly. The sizes progress as my confidence in the program increases and work out the bugs.
There is a lot of room for growth in the functionality of this project. Consequently, even though the project is listed as a 'completed' Hackster project, I will continue to add to the project and data moshing functions. There seems to be an almost endless amount of ways to produce interesting photos, and I plan to explore as many as possible within my interest and capabilities. If you want to see more of the photos that this project has produced, check out the portfolio site linked below. There are plenty of cool photos that I would like to share so I plan on making a favorites section on the site where all of my favorite data moshed photos will be posted. Feel free to use any part of this project as long as you credit it. Let me know what you think and what I should add next to this project in the comments below. Thanks for reading!
Portfolio: https://sites.google.com/view/chicagodatavisualization/home
Comments
Please log in or sign up to comment.