Documente Academic
Documente Profesional
Documente Cultură
CHAPTER 1
INTRODUCTION
The automated 3D reconstruction of neurons in brain EM image stacks remains
one of the most challenging problems in neuroscience. In such problems, neurons
spanning multiple adjacent image slices are expected to be consistently identified and
reconstructed. Conventionally, this problem has been approached as a 2D prediction
task, where each image slice is segmented individually and then, a post-processing
step was performed to generate 3D segmentation. The post-processing step usually
involves heuristic off-the-shelf classifiers that were trained to link similar segments
together across the entire image stack. These classifiers usually rely on hand-crafted
features which incorporate prior knowledge and understanding of the data. Thus,
classifiers that worked well on some problems/ datasets are not guaranteed to perform
similarly in different scenarios. It is thus desirable to design a fully trainable system
with minimal post-processing to perform the 3D segmentation task in an end-to-end
fashion. Currently, deep convolutional neural networks (CNNs) are one of the main
tools used for semantic segmentation. These models are very powerful and capable of
extracting hierarchical features from raw image data. They are characterized by their
ability to learn features directly from the raw images without relying on prior
knowledge.
CNNs have achieved success in different areas of machine learning and
computer vision. Improved performance has been achieved in image classification and
object detection tasks of real deconvolutional, and full-resolution features are not well
preserved. To address this limitation, deconvolutional networks have been proposed
recently by performing actual deconvolutional. The pooling layers are reversed in the
decoding stage by unpooling layers which keep track of the maximum activation
position selected during the pooling operation.
CNN approach attempts to design novel deep models specifically for dense
prediction problems, they do not have explicit mechanisms to address the conflicting
goals in dense prediction problems so it does not suffer from loss of information due
to excessive reduction of resolution.
2
deconvolutional layers in total as the network is not entirely symmetric. The kernel
sizes are either 3x3 or 1x1 when we performed branching before adding the residual
paths.
Zero padding was used whenever size preserving was needed in the learned
layers. They added a batch normalization layer after each learned layer and rectified
linear u units were used as the non-linearity transformation. We used a patch size of
128x128x3 in training while the entire image was used in testing. No pre-processing
was used on the raw input images. However, we modified the training labels to reduce
the segment sizes by increasing the border width in-between them. The widening of
borders was done using a minimum kernel of size 5x5 by assuming that all segments
are having a label of 1 and borders are having a label of 0 in the ground truth label
stack.
Any pixel that was in a neighbourhood of size 5x5 of a border pixel was
considered to be border as well. Label widening was crucial in allowing the network
to differentiate border from non-border pixels. This model implementation was based
on the publicly available MATLAB. They trained our RDN using back propagation
with stochastic gradient decent. The mini batch size used was 15 as the dense
prediction requires a lot of memory. However, the network requires roughly 15k
iterations to achieve its full potential due to the existence of residual paths which
speeds up the computations.
To improve the robustness of the resulting probability maps, we applied 8
variations to the testing images before passing them down through the network. A
reverse transformation was then applied to each resulting probability map before
taking the average across all variations. The transformations were combinations of
horizontal and vertical mirroring, and/or rotations by +90, -90 and 180 degrees.
1.4 EM IMAGE DENSE PREDICTION PROBLEM
In this experiments, we used two separate datasets for training and testing from
the ISBI 2013 challenge. Each dataset is a 3D stack of 100 sections from a serial
section scanning electron microscopy (SEM) of mouse cortex. The pixel resolution is
6x6x30 nm/pixel which covers a micro cube of approximately 6x6x3microns. Both
datasets have high x- and y-direction resolution whereas the resolution of z-direction
5
is low. The neurites in the training stack have been manually delineated, generating a
corresponding label stack of 100 sections. The training stack contains 400 neurites
that have been labeled consistently across the 100 slices. Some neurites are split into
several segments in some slices while still required to preserve their unique label
across sections, which increases the complexity of the 3D segmentation task. The
labels of the testing stack are not available to challenge participants.
They formulated the 3D segmentation problem of 400 neurites in the training
stack as a single 2D segmentation problem. We built a pixel classifier that accepts
patches extracted from the raw input image to generate 2D probability maps. Each
resulting probability map indicates the probability of each pixel being either a
membrane (border) pixel or non-membrane (neurite). The probability maps have no
reference of which neurite a pixel belongs to, had it been identified as a non-
membrane pixel. The final 3D segmentation was obtained.
1.5 POST PROCESSING
Super-pixel level algorithms are commonly used as a building block in most
post-processing techniques for 2D and 3D segmentation tasks. They are used mainly
to generate an over-segmentation from probability maps or affinity graphs. Later,
another classifier is built on top of the results of the super-pixel level algorithms to
accurately merge some of the overly segmented regions. The key limitation of these
approaches is that they reduce the generality of the overall proposed techniques, since
they rely on hand-crafted features to build classifiers on top of super-pixel algorithms.
One of the fundamental advantages of the proposed method is the ability to learn
features from the data, hence their ability to generalize to many other datasets.It has
been shown before that relying heavily on the learned network while simplifying post-
processing could result in a dramatic increase in the speed of computations.While
maintaining the generalization of the proposed technique.
This model tends to reduce the predicted segment sizes due to the widening of
training label borders described. As a result, we applied a reverse transformation
which used a maximum kernel to increase segments sizes. The overall processing is
fast simple and minimal additional computation is required.
6
CHAPTER 2
LITERATURE SURVEY
2.1 LITERATURE REVIEW-1: HYPERCOLUMNS FOR OBJECT
SEGMENTATION AND FINE-GRAINED LOCALIZATION, published by B.
Hariharan, P. Arbelaez, and R. Girshick, in IEEE Conf. Computer Vision
Pattern Recognition (CVPR), in the year 2015.
Recognition algorithms based on convolutional networks (CNNs) typically use the
output of the last layer as a feature representation. The methods/techniques used is
convolutional networks (CNN).
MERITS
The high resolution of the lower layers is obtained instead of up sample the
higher layers .
DEMERITS
The first problem is simultaneous detection and segmentation (SDS).
The Second problem deals with detecting an object and localizing its parts.
2.2 LITERATURE REVIEW-2: AN INTEGRATED REGION, BOUNDARY,
SHAPE- BASED ACTIVE CONTOUR FOR MULTIPLE OBJECT OVERLAP
RESOLUTION IN HISTOLOGICAL IMAGERY published by S. ALI AND A.
MADABHUSHI in the year 2011.
Synergistic boundary and region-based active contour model that incorporates shape
priors in a level set formulation with automated initialization based on watershed. The
methods/techniques used is Principle Component Analysis(PCA) .
MERITS
Segments all the overlapping and non overlapping objects within an image
simultaneously.
DEMERITS
Variability in digital slide appearance due to inconsistencies in histological
staining,
Poor image quality with tissue samples due to slide digitization
Tissue fixation.
7
An efficient LFT extraction algorithm based on image shifting and image integral. For
the splitting part, given a connected component of the segmentation map, they initially
differentiate whether it is a touching-cell clump or a single non touching cell. The
methods/techniques used is local fourier transform.
MERITS
The advantage of LFT segmentation over the others cannot be fully reflected
by the marginal improvement if only based on the segmentation accuracy.
DEMERITS
Contours containing multiple overlapping objects pose.
8
CHAPTER 3
SYSTEM METHODOLOGY
Systems methodology is the process of defining how the
information system should be built (i.e., physical system design), ensuring that the
information system is operational and used, ensuring theinformation system meets
quality standard .
3.1 METHODOLOGY OF SYSTEM IMPLEMENTATION
The first stage of any vision system is the image acquisition stage.After the image has
been obtained, various methods of processing can be applied to the image to perform
the many different vision tasks required .However, if the image has not been acquired
satisfactorily then the intended tasks may not be achievable, even with the aid of some
form of image enhancement.
where one should analyze the correlation statistics of the preceding layer and cluster
them into groups of units with high correlation. In the layers close to the input
correlated units would concentrate in local regions. These reduction layers also
include the use of rectified linear activation which makes them dual-purpose by
increasing their representation power.
3.1.5 DECONVOLUTION
Image acquired on a digital fluorescence microscope can be deconvolved, and
several new applications are being developed that apply deconvolution techniques to
transmitted light images collected under a variety of contrast enhancing strategies.
Among the most suitable subjects for improvement by deconvolution are three-
dimensional montages constructed from a series of optical sections. The basic
concepts surrounding acquisition of serial optical sections for deconvolution analysis
are presented. During deconvolution analysis, the entire series of optical sections is
analyzed to create a three-dimensional montage.
As a technique, deconvolution is often suggested as a good alternative to the
confocal microscope. This is not strictly true because images acquired using a pinhole
aperture in a confocal microscope can also be analyzed by deconvolution techniques.
3.2 SOURCES OF IMAGE DEGRADATION
The sources of image degradation can be divided into four independent
phenomena: noise, scatter, glare, and blur. The principal task of deconvolution
methods is to remove the out-of-focus blur from images. Deconvolution algorithms
can and do remove noise, but this is a relatively simple aspect of their general overall
performance.
Blur is described by a non random spreading of light that occurs by passage
through the imaging system. The most significant source of blur is diffraction, and an
image whose resolution is limited only by blur is considered to be diffraction-limited.
This represents an intrinsic limit of any imaging system and is the determining factor
in assessing the resolution limit of an optical system. Optical theory proposes
sophisticated models of blur, which can be applied, with the assistance of modern
high-speed computers, to digital images captured in the optical microscope. This is the
basis for deconvolution. It is precisely this independence of optical blur from other
11
the levels of features can be enriched by the number of stacked layers (depth).
Recent evidence reveals that network depth is of crucial importance, and the leading
results on the challenging Image Net dataset all exploit very deep modes, with a
depth of sixteen to thirty.
3.5 IMAGE RESOLUTION
3.6.1Pixel resolution
3.6.2Spatial resolution
3.6.3Spectral resolution
3.6.4Temporal resolution
3.6.5Radiometric resolution
same standards, the "Number of Effective Pixels" that an image sensor or digital
camera has is the count of pixel sensors that contribute to the final image (including
pixels not in said image but nevertheless support the image filtering process), as
opposed to the number of total pixels, which includes unused or light-shielded pixels
around the edges.The count of pixels is not a real measure of the resolution of digital
camera images, because color image sensors are typically set up to alternate color
filter types over the light sensitive individual pixel sensors. Digital images ultimately
require a red, green, and blue value for each pixel to be displayed or printed, but one
individual pixel in the image sensor will only supply one of those three pieces of
information. The image has to be interpolated or decimated to produce all three colors
for each output pixel.
The measure of how closely lines can be resolved in an image is called spatial
resolution, and it depends on properties of the system creating the image, not just the
pixel resolution in pixels per inch. For practical purposes the clarity of the image is
decided by its spatial resolution, not the number of pixels in an image. In effect,
spatial resolution refers to the number of independent pixel values per unit length.
Spectral resolution is the ability to resolve spectral features and bands into their
separate components. Color images distinguish light of different spectra. Multispectral
images resolve even finer differences of spectrum or wavelength than is needed to
reproduce color. That is, multispectral images have higher spectral resolution than
normal color images. Spectral resolution is the ability to resolve spectral features and
bands into their separate components. The spectral resolution required by the analyst
or researcher depends upon the application involved.
of Fourier transform. The time resolution used for movies is usually 24 to 48 frames
per second (frames/s).
3.6.5 RADIOMETRIC RESOLUTION
Radiometric resolution determines how finely a system can represent or
distinguish differences of intensity, and is usually expressed as a number of levels or a
number of bits, for example 8 bits or 256 levels that is typical of computer image files.
The higher the radiometric resolution, the better subtle differences of intensity
or reflectivity can be represented, at least in theory. In practice, the effective
radiometric resolution is typically limited by the noise level, rather than by the number
of bits of representation.
Electron microscopy is the most thoroughly developed approach for the dense
reconstruction of neural circuits. Because the wavelength of an electron under 10 kV
accelerating voltage is 10 pm, imaging with electrons can (in principle) reach spatial
resolutions in the sub-nanometer to nanometer range, more than sufficient to trace the
finest morphological sub-structures of neurons. The basic strategy employed by the
current EM approaches is to obtain many morphological images of thin tissue
sections, segmenting those images into regions corresponding to distinct neuronal
processes, and tracing individual axons from one image to another. Because axons are
thin, long, and densely interspersed with other neuronal processes, tracing their entire
lengths is a challenge.
Merkle used the number of electrons per voxel, the number of parallel electron
microscopes available, and the total project time to estimate the beam current per
microscope: imaging a whole human brain in 3 years at 10nm10nm10nm voxel
size, with 7-bit precision and 1000 parallel microscopes, would give 0.1 mA beam
current, comparable with that of electron microscopes circa 1989.
CHAPTER 4
SOFTWARE DESCRIPTION
4.1 GENERAL
In 2004, MATLAB had around one million users across industry and
academia. MATLAB users come from various backgrounds of engineering, science,
and economics.
Built-in graphics for visualizing data and tools for creating custom plots
Libraries written in Perl, Java, ActiveX or .NET can be directly called from
MATLAB, and many MATLAB libraries (for example XML or SQL support) are
implemented as wrappers around Java or ActiveX libraries. Calling MATLAB from
Java is more complicated, but can be done with a MATLAB toolbox which is sold
separately by MathWorks, or using an undocumented mechanism called JMI (Java-to-
MATLAB Interface), (which should not be confused with the unrelated Java Metadata
Interface that is also called JMI). Official MATLAB API for Java was added in 2016.
Development environment
Spreadsheet import tool that provides more options for selecting and loading
mixed textual and numeric data.
Readability and navigation improvements to warning and error message in the
MATLAB command window.
Automatic variable and function renaming in the MATLAB editor.
MATLAB provides a high level language development tools that let you
quickly develop and analyze your algorithms and applications.
The MATLAB language provides native support for the vector and matrix
operations that are fundamental to solving engineering and scientific problems,
MATLAB Editor
Code Analyzer
The Code Analyzer checks your code for problems and recommends
modifications to maximize performance and maintainability.
MATLAB Profiler
Directory Reports
Scan all the files in the directory and report on the code efficiency, file
differences, file dependencies and code coverage.
MATLAB supports the entire data analysis process, from acquiring data from
external device and database, through preprocessing, visualization, and numerical
analysis, to producing presentation-quality output.
Data Analysis
Data access
ADVANTAGE
It uses a large amount of memory and on slow computers it is very hard to use.
It sits on top of Windows, getting as much CPU time as Windows allows it to
have. This makes real-time applications very complicated.
24
CHAPTER 5
RESULTS
In this, the proposed model is very effective.Achieving the conflicting
goals in dense output prediction.Full-resolution predictions and including
sufficient contextual information.
5.1 Input Image
The above figure shows the RAW image of Electron Microscopy image Segmentation
which has 1024x768 pixels
25
The above figure shows the Colour Map Generation image. This uses
the color function to display a 16-by-16 matrix whose elements vary from 0 to
255 in a rectilinear spiral. The hsv color map starts with red in the center, then
passes through yellow, green, cyan, blue, and magenta before returning to red
at the outside end of the spiral. A color map is an m-by-3 matrix of real
numbers between 0.0 and 1.0. Each row is an RGB vector that defines one
color. The kth row of the color map defines the kth color,
where map(k,:) = [r(k) g(k) b(k)]) specifies the intensity of red, green, and blue.
26
The above figure shows the Identification of electron microscopy disease using signal
image segmentation specific for one-dimensional signals and does not trivially
transfer to higher-dimensional images. Nevertheless, this general idea has inspired
several other authors to investigate coarse-to-fine schemes for image segmentation.
28
The above figure shows the Identification of electron microscopy disease using color
image EM image segmentation MRI image.
29
CHAPTER 6
CONCLUSION
We proposed a computational technique for EM image segmentation by
obtaining dense predictions that combined multi-scale contextual reasoning along with
full-resolution reconstruction. This approach achieved promising performance while
relying on minimum post-processing. They expect better probability maps be
generated with improvement in the z-dimension resolution of the data provided. A
limitation in the underlying post-processing techniques is that it requires a specific
parameter to control the level of over/under segmentation. Automatic tuning of this
parameter is not straightforward and can be data-specific even if it is tuned on the
validation dataset.They used semi automated visualization of the segmentations to
overcome this limitation. Nonetheless, our method can be paired with any other post-
processing techniques, leading to an overall performance improvement.
They did not use hand-crafted features either in the network training or post-
processing stages. Consequently, we demonstrated the ability of this model to
generalize by applying it to multiple datasets obtained from different species. This
method achieved consistently promising performance.They believe this method can
generalize well to other similar dense output prediction task.
30
REFERENCES