May 2020 edit: updated package versions and made small adjustments to the code.
Ever tried to visualize 3D images using Python? It sounds like something that would come up frequently when using things like medical scanner data, but it’s not super well documented. One way to go about it is display 2D slides, possibly interactively. DataCamp has a good tutorial on how to do this, but what if you can’t use a dynamic image? For instance, for a printed publication, a static image is your only option.
For this tutorial, you’ll need the requests library to get the data, nibabel to read the images, numpy and scikit-image for various manipulation tasks, and of course matplotlib for the actual plotting. You can get all of this by running
pip install requests nibabel numpy scikit-image matplotlib
Note that I’m using Python 3 here (as you should be), but this tutorial should work with minor changes using Python 2. You can check the library versions I’ve used at the very end of the page.
You can also have a peek at the results before embarking.
Getting the data
We’ll use the Attention to Visual Motion fMRI dataset1Büchel, Christian, and K. J. Friston. “Modulation of connectivity in visual pathways by attention: cortical interactions evaluated with structural equation modelling and fMRI.” Cerebral cortex (New York, NY: 1991) 7.8 (1997): 768-778. for cool brain images. Of course, feel free to visualize whatever you want. As long as you have a 3D numpy array of data, you can skip to the next section to get started on the actual visualizing.
Let’s start by getting the zip file (79MB):
import requests
images = requests.get('http://www.fil.ion.ucl.ac.uk/spm/download/data/attention/attention.zip')
To avoid cluttering the filesystem, let’s keep the zip archive in memory. We’re using BytesIO
, which, like its cousin StringIO
, is a essentially a way to equip a bytes
(or string
) object with file I/O operations (such as read
, write
and seek
).
import zipfile
from io import BytesIO
zipstream = BytesIO(images.content)
zf = zipfile.ZipFile(zipstream)
This zip file contains a number of images in Analyze format. It’s easy to open an image with nibabel: just run nibabel.load(filename)
. Unfortunately, a single Analyze-formatted image consists of a header file (.hdr
) and a separate file for the data itself (.img
). If the images are stored on disk, nibabel.load
will automatically find both files, but this doesn’t work here.
With a little trickery, though, we can get nibabel
to load the image directly from memory. img.get_data()
gets us the 3D data array, and we can get started with plotting!
from nibabel import FileHolder
from nibabel.analyze import AnalyzeImage
header = BytesIO(zf.open('attention/structural/nsM00587_0002.hdr').read())
image = BytesIO(zf.open('attention/structural/nsM00587_0002.img').read())
img = AnalyzeImage.from_file_map({'header': FileHolder(fileobj=header), 'image': FileHolder(fileobj=image)})
arr = img.get_fdata()
arr.shape
(157, 189, 68)
A first look
Let’s first do a regular plot of a horizontal slice of our brain:
import matplotlib.pyplot as plt
%matplotlib inline
plt.imshow(arr[:,:,5])
plt.colorbar()
plt.plot()
[]
My knowledge of neurology is severely limited, but I’m pretty sure those are eye sockets on the right!
If you look at the colorbar, you’ll realize that there are very few points that reach the top values (larger than, say, 1500). We can confirm that insight by making a histogram:
(BTW, we’re using the default Viridis colorscheme, which was designed to solve a number of ergonomic problems with Jet, the former default. There’s an interesting talk about its design process (video).)
import numpy as np
from matplotlib import cm
def normalize(arr):
arr_min = np.min(arr)
return (arr-arr_min)/(np.max(arr)-arr_min)
def show_histogram(values):
n, bins, patches = plt.hist(values.reshape(-1), 50, density=True)
bin_centers = 0.5 * (bins[:-1] + bins[1:])
for c, p in zip(normalize(bin_centers), patches):
plt.setp(p, 'facecolor', cm.viridis(c))
plt.show()
show_histogram(arr)
I’m going to change this using a homebrewed and totally guesstimated transformation:
def scale_by(arr, fac):
mean = np.mean(arr)
return (arr-mean)*fac + mean
transformed = np.clip(scale_by(np.clip(normalize(arr)-0.1, 0, 1)**0.4, 2)-0.1, 0, 1)
show_histogram(transformed)
The range of colors is more uniform, with a big peak at 0 (which we’ll see is a good idea).
Plotting!
Warmup: a pile of cubes
Matplotlib’s 3D capabilities are still being developed, and they have a few quirks we’ll have to work around. Let’s forget our brain for a moment, and start with a very simple voxel plot, to introduce basic concepts.
from mpl_toolkits.mplot3d import Axes3D
def make_ax(grid=False):
fig = plt.figure()
ax = fig.gca(projection='3d')
ax.set_xlabel("x")
ax.set_ylabel("y")
ax.set_zlabel("z")
ax.grid(grid)
return ax
filled = np.array([
[[1, 0, 1], [0, 0, 1], [0, 1, 0]],
[[0, 1, 1], [1, 0, 0], [1, 0, 1]],
[[1, 1, 0], [1, 1, 1], [0, 0, 0]]
])
ax = make_ax(True)
ax.voxels(filled, edgecolors='gray', shade=False)
plt.show()
So we can display some voxels easily enough (in case you don’t know what they are and for some reason haven’t clicked the Wikipedia link above, they’re just 3D pixels, so little cubes).
Here, the filled
array tells matplotlib which voxels to fill in. Any truthy value (you’ll most likely use True
or 1
) in the array means that the voxel in the corresponding poxition should be filled. It’s in x
-major, z
-minor order, so the first [1, 0, 1]
block, for instance, means “fill in the blocks at (x, y, z)=(0, 0, 0) don’t fill the one at (0, 0, 1), and fill the one at (0, 0, 2)”.
That’s all good, but the blocks in front tend to occlude the rest. How about we make everything semi-transparent? We can do this by specifying an HTML color with an alpha component.
ax = make_ax()
ax.voxels(filled, facecolors='#1f77b430', edgecolors='gray', shade=False)
plt.show()
That seems to work (although it’s a bit messy), but there’s a problem. Look at what happens if we display a solid block:
ax = make_ax()
ax.voxels(np.ones((3, 3, 3)), facecolors='#1f77b430', edgecolors='gray', shade=False)
plt.show()
Only the faces and edges on the sides are rendered. This seems okay, but what if we want to color the middle voxel red, it’s not going to show:
ax = make_ax()
colors = np.array([[['#1f77b430']*3]*3]*3)
colors[1,1,1] = '#ff0000ff'
ax.voxels(np.ones((3, 3, 3)), facecolors=colors, edgecolor='gray', shade=False)
plt.show()
To solve this, we can add extra voxels in between. Let’s define a helper function, explode
, which will take our filled
array and return an array twice as large in each dimension, with an extra space between each voxel:
def explode(data):
shape_arr = np.array(data.shape)
size = shape_arr[:3]*2 - 1
exploded = np.zeros(np.concatenate([size, shape_arr[3:]]), dtype=data.dtype)
exploded[::2, ::2, ::2] = data
return exploded
(Note that the function supports arrays of more than three dimensions, and will stick any extra dimensions back at the end. This is because we’ll later use explode on 4D arrays.)
Let’s plot the exploded version, without the borders for better visibility:
ax = make_ax(True)
colors = np.array([[['#1f77b430']*3]*3]*3)
colors[1,1,1] = '#ff0000ff'
colors = explode(colors)
filled = explode(np.ones((3, 3, 3)))
ax.voxels(filled, facecolors=colors, shade=False)
plt.show()
This works, but we don’t want these gaps between the voxels.
Fortunately, voxels
supports one (or rather three) other arguments, x
, y
and z
. These should follow the format generated by np.indices
, so they’re 4-dimensional. They allow us to change the coordinates of a voxel corner. Specifically, x[i, j, k]
is the x coordinate of the lower-bottom-left corner of the voxel at position (i, j, k) in the filled
array. This also mean that if you have m by n by p voxels, x
, y
and z
all have shape (m+1, n+1, p+1).
Let’s briefly go back to the full block example. We can extend the middle block by incrementing y
for all blocks in the middle and back columns:
ax = make_ax()
filled = np.ones((3, 3, 3))
x, y, z = np.indices(np.array(filled.shape) + 1)
y[:,2:,:] += 4
ax.voxels(x, y, z, filled, facecolors='#1f77b430', edgecolor='gray', shade=False)
plt.show()
Or we could change the coordinates of a specific point:
ax = make_ax()
filled = np.ones((3, 3, 3))
x, y, z = np.indices(np.array(filled.shape) + 1)
x[3, 2, 2] = 5
ax.voxels(x, y, z, filled, facecolors='#1f77b430', edgecolor='gray', shade=False)
plt.show()
Putting all this together, we can draw over all the inserted voxels we just added by making all of the original voxels twice as large. Let’s do that, and plot our red cube example again:
def expand_coordinates(indices):
x, y, z = indices
x[1::2, :, :] += 1
y[:, 1::2, :] += 1
z[:, :, 1::2] += 1
return x, y, z
ax = make_ax()
colors = np.array([[['#1f77b430']*3]*3]*3)
colors[1,1,1] = '#ff0000ff'
colors = explode(colors)
filled = explode(np.ones((3, 3, 3)))
x, y, z = expand_coordinates(np.indices(np.array(filled.shape) + 1))
ax.voxels(x, y, z, filled, facecolors=colors, edgecolors='gray', shade=False)
plt.show()
Serious business
An important note: voxels
is slow on large 3D data. I’ve resized all the brain data to be 50×50×50 voxels, which still has a full runtime of nearly 4 minutes for the entire skull (on my admittedly not very beefy machine). If you have a powerful machine, you can set a higher image size, but remember that it scales cubically. I’m resizing using scikit-image, but there are lots of other options.
(Remember that transformed
is our 3D data, with rescaled values.)
IMG_DIM = 50
from skimage.transform import resize
resized = resize(transformed, (IMG_DIM, IMG_DIM, IMG_DIM), mode='constant')
We’ve dealt with binary data so far, but how can we represent 3D volumetric data with a range of values? The approach I’ve taken is to set each voxel’s transparency equal to its value. In the case of brain data, this allows to see through the black areas corresponding to air around the head, as well as through some of the empty parts inside the brain. Of course, you can’t see much trough the brain itself—but the image would be incredibly confusing otherwise. In any case, the technique I’m showing here allows you to adapt the color/transparency system easily should you want to implement something else (for instance, you could set the maximal transparency to a low value, which would make it possible to see through your plot much more).
I’ve replicated explode
and expand_coordinates
from above, so you can directly copy-paste the cell below to use in your project.
def explode(data):
shape_arr = np.array(data.shape)
size = shape_arr[:3]*2 - 1
exploded = np.zeros(np.concatenate([size, shape_arr[3:]]), dtype=data.dtype)
exploded[::2, ::2, ::2] = data
return exploded
def expand_coordinates(indices):
x, y, z = indices
x[1::2, :, :] += 1
y[:, 1::2, :] += 1
z[:, :, 1::2] += 1
return x, y, z
def plot_cube(cube, angle=320):
cube = normalize(cube)
facecolors = cm.viridis(cube)
facecolors[:,:,:,-1] = cube
facecolors = explode(facecolors)
filled = facecolors[:,:,:,-1] != 0
x, y, z = expand_coordinates(np.indices(np.array(filled.shape) + 1))
fig = plt.figure(figsize=(30/2.54, 30/2.54))
ax = fig.gca(projection='3d')
ax.view_init(30, angle)
ax.set_xlim(right=IMG_DIM*2)
ax.set_ylim(top=IMG_DIM*2)
ax.set_zlim(top=IMG_DIM*2)
ax.voxels(x, y, z, filled, facecolors=facecolors, shade=False)
plt.show()
A few notes on this implementation:
- Instead of colors as strings, I’m using a 4D colors array, where the last dimension (of size 4) holds the red, green, blue, and alpha (transparency) values. Doing
facecolors[:,:,:,-1] = cube
makes the alpha equal to the voxel value. - I’m still using Viridis, the default color map. You can use any map you like that’s supported by matplotlib, by changing the call to
cm.viridis
. - I’m setting some axis limits to make sure that all the plots are on the same scales, even if I truncate the image to show a cross-section.
- You can add a call to
ax.set_axis_off()
if you want to remove the background and axis ticks.
Oh, and if you were wondering, this is where explode
handling 4D arrays comes in handy.
Results
So first, a cut view of the skull:
plot_cube(resized[:35,::-1,:25])
(I’m plotting the y-axis backwards so that the eyes are in front).
A view from the back, cutting through in diagonal:
cube = np.copy(resized)
for x in range(0, IMG_DIM):
for y in range(0, IMG_DIM):
for z in range(max(x-y+5, 0), IMG_DIM):
cube[x, y, z] = 0
plot_cube(cube, angle=200)
And a full view, where you can see the nasal cavity and make out the eye sockets at the very bottom:
plot_cube(resized[:,::-1,:])
I hope you enjoyed this tutorial! Feel free to drop me suggestions for improvements, questions, or other random notes below. You can also look at voxel
’s documentation for more details.
Here are the library versions I’ve used for this tutorial:
matplotlib==3.2.1
nibabel==3.1.0
numpy==1.18.4
requests==2.23.0
scikit-image==0.17.2
You can also download the notebook.
-
Büchel, Christian, and K. J. Friston. “Modulation of connectivity in visual pathways by attention: cortical interactions evaluated with structural equation modelling and fMRI.” Cerebral cortex (New York, NY: 1991) 7.8 (1997): 768-778. ↩
Comments