Analyzing huge pathology images with open source software
© Deroulers et al.; licensee BioMed Central Ltd. 2013
Received: 2 May 2013
Accepted: 22 May 2013
Published: 6 June 2013
Digital pathology images are increasingly used both for diagnosis and research, because slide scanners are nowadays broadly available and because the quantitative study of these images yields new insights in systems biology. However, such virtual slides build up a technical challenge since the images occupy often several gigabytes and cannot be fully opened in a computer’s memory. Moreover, there is no standard format. Therefore, most common open source tools such as ImageJ fail at treating them, and the others require expensive hardware while still being prohibitively slow.
We have developed several cross-platform open source software tools to overcome these limitations. The NDPITools provide a way to transform microscopy images initially in the loosely supported NDPI format into one or several standard TIFF files, and to create mosaics (division of huge images into small ones, with or without overlap) in various TIFF and JPEG formats. They can be driven through ImageJ plugins. The LargeTIFFTools achieve similar functionality for huge TIFF images which do not fit into RAM. We test the performance of these tools on several digital slides and compare them, when applicable, to standard software. A statistical study of the cells in a tissue sample from an oligodendroglioma was performed on an average laptop computer to demonstrate the efficiency of the tools.
Our open source software enables dealing with huge images with standard software on average computers. They are cross-platform, independent of proprietary libraries and very modular, allowing them to be used in other open source projects. They have excellent performance in terms of execution speed and RAM requirements. They open promising perspectives both to the clinician who wants to study a single slide and to the research team or data centre who do image analysis of many slides on a computer cluster.
The virtual slide(s) for this article can be found here:
KeywordsDigital pathology Image processing Virtual slides Systems biology ImageJ NDPI
Virtual microscopy has become routinely used over the last few years for the transmission of pathology images (the so-called virtual slides), for both telepathology and teaching[1, 2]. In more and more hospitals, virtual slides are even attached to the patient’s file[3, 4]. They have also a great potential for research, especially in the context of multidisciplinary projects involving e.g. mathematicians and clinicians who do not work at the same location. Quantitative histology is a promising new field, involving computer-based morphometry or statistical analysis of tissues[5–9]. A growing number of works report the pertinence of such images for diagnosis and classification of diseases, e.g. tumours[10–14]. Databases of clinical cases will include more and more digitized tissue images. This growing use of virtual microscopy is accompanied by the development of integrated image analysis systems offering both virtual slide scanning and automatic image analysis, which makes integration into the daily practice of pathologists easier. See Ref. for a review of some of these systems.
However, this technique involves the manipulation of huge images (of the order of 10 billions of pixels for a full-size slide at magnification 40x with a single focus level) for which the approach taken by most standard software, loading and decompressing the full image into RAM, is impossible (a single slice of a full-size slide needs of the order of 30 GiB of RAM). As a result, standard open-source software such as ImageJ, ImageMagick or GraphicsMagick completely fails or is prohibitively slow when used on these images. Of course, commercially available software exists, but it is usually quite expensive, and very often restricted to a single operating system. It uses proprietary source code, which is a problem if one wants to control or check the algorithms and their parameters when doing image analysis for research.
In addition, many automated microscopes or slide scanners store the images which they produce into proprietary or poorly documented file formats, and the software provided by vendors is often specific to some operating system. This leads to several concerns. First, it makes research based on digital pathology technically more difficult. Even when a project is led on a single site, one has often to use clusters of computers to achieve large-scale studies of many full-size slides from several patients. Since clusters of computers are typically run by open source software such as Linux, pathology images stored in non-standard file formats are a problem. Furthermore, research projects are now commonly performed in parallel in several sites, not to say in several countries, thanks to technology such as Grid, and there is ongoing efforts towards the interoperability of information systems used in pathology[3, 22]. Second, proprietary formats may hinder the development of shared clinical databases and access of the general public to knowledge, whereas the citizen should receive benefit of public investments. Finally, they may also raise financial concerns and conflicts of interest.
There have been recent attempts to define open, documented, vendor-independent software[24, 25], which partly address this problem. However, very large images stored in the NDPI file format produced by some slide scanners manufactured by Hamamatsu, such as the NanoZoomer, are not yet fully supported by such software. For instance, LOCI Bio-Formats is presently unable to open images, one dimension of which is larger than 65k, and does not deal properly with NDPI files of more than 4 GiB. OpenSlide does not currently support the NDPI format. NDPI-Splitter needs to be run on Windows and depends on a proprietary library.
To address these problems, we have developed open source tools which achieve two main goals: reading and converting images in the NDPI file format into standard open formats such as TIFF, and splitting a huge image, without decompressing it entirely into RAM, into a mosaic of much smaller pieces (tiles), each of which can be easily opened or processed by standard software. All this is realized with high treatment speed on all platforms.
The main software is implemented in the C programming language as separate, command-line driven executables. It is independent of any proprietary library. This ensures portability on a large number of platforms (we have tested several versions of Mac OS X, Linux and Windows), modularity and ease of integration into scripts or other software projects.
It is complemented by a set of plugins for the public domain software ImageJ, implemented in Java, which call the main executables in an automatic way to enable an interactive use.
The LargeTIFFTools and NDPITools are based on the open source TIFF and JPEG or libjpeg-turbo libraries. The NDPITools plugins for ImageJ are based on the Java API of ImageJ[17, 30] and on the open source software Image-IO, and use the Java Advanced Imaging 1.1.3 library.
splitting a tiled TIFF file into multiple TIFF files, one for each of the tiles (tiffsplittiles program);
extracting (“cropping”) quickly a given rectangle of a supposedly tiled TIFF file into a TIFF or JPEG file (tifffastcrop program);
splitting one or several TIFF file(s), possibly very large, into mosaic(s), without fully decompressing them in memory (tiffmakemosaic program);
converting a NDPI file into a standard multiple-image TIFF file, tiled if necessary, using upon request the BigTIFF format introduced in version 4.0.0 of the TIFF library [27, 33, 34], and encoding magnification and focus levels as TIFF “image description” fields (ndpi2tiff program);
creating a standard TIFF file for all or part of the magnification levels and focus levels present in the given NDPI file (the user can ask for specific magnification and focus levels and for a specific rectangular region of the image), and, upon request, creating a mosaic for each image which doesn’t fit into RAM or for all images (ndpisplit program). The names of the created files are built on the name of the source file and incorporate the magnification and focus levels (and, in the case of mosaic pieces, the coordinates inside the mosaic).
A mosaic is a set of TIFF or JPEG files (the pieces) which would reproduce the original image if reassembled together, but of manageable size by standard software. The user can either specify the maximum amount of RAM which a mosaic piece should need to be uncompressed (default: 1024 MiB), or directly specify the size of each piece. In the first case, the size of each piece is determined by the software. A given amount of overlap between mosaic pieces can be requested, either in pixels or as a percentage of the image size. This is useful e.g. for cell counting, not to miss cells which lie on the limit between two adjacent pieces.
Our tools can be used through the command line (POSIX-like shell or Windows command interpreter), and therefore can be very easily integrated into scripts or other programs. Depending on the tool, the paths and file names of one or several files, in NDPI or TIFF format, have to be provided. Options can be added with their arguments on the command line to modify the behavior of the programs from its default. They are explained in the messages printed by the programs run without arguments, in Unix-style man pages, and on the web pages of the project (see below in the Availability and requirements Section).
Under the Windows OS, one can click-and-drag the NDPI file icon onto the icon of ndpi2tiff or ndpisplit. We provide precompiled binaries where frequently-used options are turned on by default: e.g. ndpisplit-mJ.exe produces a mosaic in JPEG format as with option -mJ. The conversion result or mosaic can be found in the same directory as the original NDPI image.
Results and discussion
We compare the performance of our tools on several fundamental tasks to standard, broadly available software in representative examples and on broadly available computers.
Making a mosaic from a huge image
We chose an 8-bit RGB colour JPEG-compressed TIFF file of 103168 ×63232 pixels originating in the digitization of a pathology slide. The original file weighted 975.01 MiB. Loading this image entirely into RAM would need at least 3×103168×63232=18.2 GiB and is presently intractable on most if not all desktop and laptop computers of reasonable cost.
The task was to produce, from this image, a mosaic of 64 pieces so that each one needs less than 512 MiB RAM to open.
On a 3.2 GHz Intel Core i3 IMac computer with 16 GB of RAM, the convert command from ImageMagick (version 6.8.0-7 with quantum size 8 bits) was unable to complete the request. GraphicsMagick (gm convert-crop; version 1.3.17 with quantum size 8 bits) completed the request in 70 min, using 25 GiB of disk space. tiffmakemosaic from our LargeTIFFTools completed the request in 2.5 min.
To ascertain that this task can be equally achieved even on computers with a modest RAM amount, we performed the same task on a 6-year-old 2.66 GHz Core2Duo Intel IMac with 2 GiB RAM. The task was completed in 9.0 min.
Converting NDPI into TIFF
Splitting a NDPI file into TIFF files. A pathology sample (6.7 cm2 of tissue) was scanned at magnification 40x and with 11 focus levels (every 2 microns) by a NanoZoomer, resulting in a 6.5 GiB file in proprietary NDPI format (called file a.ndpi hereafter). On a 2.6 GHz Intel Core i7 Mac Mini computer with 16 GiB RAM, ndpisplit extracted all 55 images (11 focus levels and 5 magnifications) as independent, single-image TIFF files with JPEG compression in 7.11 min. The size of the largest images was 180224×70144. The speed was limited only by the rate of I/O transfers since the CPU usage of this task was 1.38 min, out of which the system used 1.30 min. Executing again the same task straight after the first execution took only 0.57 min because the NDPI file was still in the cache of the operating system.
To ascertain that this task can be equally achieved even on computers with a modest RAM amount, we made a try on a 6-year-old 2.66 GHz Core2Duo Intel PC with 2 GiB RAM running 32-bits Windows XP Pro SP3. The original file shown in Figure1, called b.ndpi, and weighting 2.07 GiB (largest image: 103168×63232 pixels), was split into independent TIFF files in 2.2 min without swapping.
In comparison, the LOCI Bio-Formats plugins for ImageJ, in its version 4.4.6 with ImageJ 1.43 m, was not able to open the images in file a.ndpi even at low resolution.
Converting a NDPI file into a multiple-images TIFF file. Alternatively, the same proprietary-format file a.ndpi was converted into a multiple-images TIFF file with ndpi2tiff. On the same computer as before, the conversion time was 7.0 min. Here again, the speed of the process is limited only by the rate of I/O transfers since the conversion took only 30 s if performed when the NDPI file was still in the cache of the operating system.
Since the resulting TIFF file could not store all 55 images in less than 4 GiB, we passed the option -8 on the command line to ndpi2tiff to request using the BigTIFF format extension. The specifications of this extension to the TIFF standard, discussed and published before 2008[33, 34], are supported by LibTIFF as of version 4.0.0, and therefore by the abundant image viewing and manipulation software which relies on LibTIFF. If the use of the BigTIFF format extension would have impeded the further exploitation of the produced TIFF file, we could have simply used ndpisplit as above. Or we could have called the ndpi2tiff command several times, each time requesting extraction of a subset of all images by specifying image numbers after the file name, separated with commas, as in a.ndpi,0,1,2,3,4.
Extracting a small region from a huge image
This task can be useful to visualize at full resolution a region of interest which the user has selected on a low-magnification preview image. Therefore, it should be performed as quickly as possible.
From a TIFF file
Speed comparison of software to extract a 256×256 rectangle from a huge TIFF image
Image size (px)
From a NDPI file
The task was to extract a rectangular region of size 256×256 from one of the largest images of the file a.ndpi (size 180224×70144). On a 2.6 GHz Intel Core i7 Mac Mini computer with 16 GB of RAM, the execution time was 0.12 s for one extract, and in average 0.06 s per extract in a series of 20 extracts with locations drawn uniformly at random inside the whole image.
Integration in digital pathology image servers or virtual slide systems
The NDPITools are being used in several other software projects:
in WIDE, to deal with NDPI files. WIDE is an open-source biological and digital pathology image archiving and visualization system, which allows the remote user to see images stored in a remote library in a browser. In particular, thanks to the feature of high-speed extraction of a rectangular region by ndpisplit, WIDE saves costly disk space since it doesn’t need to store TIFF files converted from NDPI files in addition to the latter.
Exploiting a large set of digital slides
In the framework of a study about invasive low-grade oligodendrogliomas reported elsewhere, we had to deal with 303 NDPI files, occupying 122 GiB. On a 3.2 GHz Intel Core i3 IMac computer with 16 GB RAM, we used ndpisplit in a batch work to convert them into standard TIFF files, which took only a few hours. The experimental -s option of ndpisplit was used to remove the blank filling between scanned regions, resulting in an important disk space saving and in smaller TIFF files (one for each scanned region) which where easier to manipulate afterwards. Then, for each sample, Preview.app and ImageJ were used to inspect the resulting images and manually select the regions of clinical interest. The corresponding extracts of the high magnification images were the subject of automated cell counting and other quantitative analyses using ImageJ. In particular, we collected quantitative data about edema or tissue hyperhydration. This quantity needed a specific image analysis procedure which is not offered by standard morphometry software and, unlike cell density estimates, could not be retrieved by sampling a few fields of view in the microscope. Therefore, virtual microscopy and our tools were essential in this study.
Study of a whole slide of brain tissue invaded by an oligodendroglioma
To demonstrate the possibility to do research on huge images even with a modest computer, we chose a 3-year-old MacBook Pro laptop computer with 2.66 GHz Intel Core 2 Duo and 4 GiB of RAM. We used ImageJ and the NDPITools to perform statistics on the upper piece of tissue on the slide shown in Figure1.
Since the digital slide b.ndpi weighted 2.07 GiB, with a high resolution image of 103168×63232 pixels, it was not possible to do the study in a straightforward way. We opened the file b.ndpi as a preview image with the command Plugins > NDPITools > Preview NDPI... and selected on it the left tissue sample. Then we used the command Plugins > NDPITools > Custom extract to TIFF / Mosaic... and asked for extraction as a mosaic of 16 JPEG files, each one needing less than 1 GiB of RAM to open, and with an overlap of 60 pixels. This was completed within a few minutes. Then we applied an ImageJ macro to each of the 16 pieces to identify the dark cell nuclei (those with high chromatin content), based on thresholding the luminosity values of the pixels, as shown in Figure1. It produced text files with the coordinates and size of each cell nucleus.
Out of the 154240 identified nuclei, 1951 were positioned on the overlapping regions between pieces. Using the overlap feature of our tools enabled to properly detect these nuclei, since they would have been cut by the boundary of the pieces of the mosaic in absence of overlap. We avoided double counting by identifying the pairs of nuclei situated in the overlapping regions and which were separated by a distance smaller than their radius.
The LargeTIFFTools, NDPITools and NDPITools plugins for ImageJ achieve efficiently some fundamental functions on large images and in particular digital slides, for which standard open source software fails or performs badly. They enable both the clinician to examine a single slide and the bioinformatics research team to perform large-scale analysis of many slides, possibly on computer grids.
Downloads of the NDPITools
Windows (32 bits)
Mac OS X
We have explained how the software was used to study some microscopic properties of brain tissue when invaded by an oligodendroglioma, and we have given an illustrative application to the analysis of a whole-size pathology slide. This suggests other promising applications.
Availability and requirements
Project name: LargeTIFFTools
Operating system(s): Platform independent
Programming language: C
Other requirements: libjpeg, libtiff
License: GNU GPLv3
Project name: NDPITools
Project home page: http://www.imnc.in2p3.fr/pagesperso/deroulers/software/ndpitools/
Operating system(s): Platform independent
Programming language: C
Other requirements: —
License: GNU GPLv3
For the convenience of users, precompiled binaries are provided for Windows (32 and 64 bits), Mac OS X and Linux.
- c.NDPITools plugins for ImageJ
Project name: NDPITools plugins for ImageJ
Project home page: http://www.imnc.in2p3.fr/pagesperso/deroulers/software/ndpitools/
Operating system(s): Platform independent
Programming language: Java
Other requirements: ImageJ 1.31s or higher, Ant, JAI 1.1.3
License: GNU GPLv3
We thank F. Bouhidel and P. Bertheau for their help with the slide scanner of the Pathology Laboratory of the Saint-Louis Hospital in Paris, and C. Klein (Imaging facility, Cordeliers Research Center – INSERM U872, Paris) for tests and suggestions.
The computer, CPU, operating system, and programming language names quoted in this article are trademarks of their respective owners.
- Diamond J, McCleary D: Virtual microscopy. Advanced Techniques in Diagnostic Cellular Pathology. Edited by: Hannon-Fletcher M, Maxwell P. 2009, Chichester UK: John Wiley & Sons, Ltd;
- Ameisen D, Yunès JB, Deroulers C, Perrier V, Bouhidel F, Battistella M, Legrès L, Janin A, Bertheau P: Stack or Trash? Fast quality assessment of virtual slides. Diagn Pathol. 2013, in press
- García Rojo M, Castro AM, Gonçalves L: COST action “EuroTelepath”: digital pathology integration in electronic health record, including primary care centres. Diagn Pathol. 2011, 6 (Suppl 1): S6-10.1186/1746-1596-6-S1-S6.PubMed CentralView Article
- Ameisen D: Intégration des lames virtuelles dans le dossier patient électronique. PhD thesis. 2013, Univ Paris Diderot-Paris 7
- Collan Y, Torkkeli T, Personen E, Jantunen E, Kosma VM: Application of morphometry in tumor pathology. Anal Quant Cytol Histol. 1987, 9 (2): 79-88.PubMed
- Wolfe P, Murphy J, McGinley J, Zhu Z, Jiang W, Gottschall E, Thompson H: Using nuclear morphometry to discriminate the tumorigenic potential of cells: A comparison of statistical methods. Cancer Epidemiol Biomarkers Prev. 2004, 13 (6): 976-988.PubMed
- Gürcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B: Histopathological image analysis: a review. Biomed Eng, IEEE Rev. 2009, 2: 147-171.View Article
- Gerin C, Pallud J, Deroulers C, Varlet P, Oppenheim C, Roux FX, Chrétien F, Thomas SR, Grammaticos B, Badoual M: Quantitative characterization of the imaging limits of diffuse low-grade oligodendrogliomas. Neuro-Oncol. 2013., in press [http://dx.doi.org/10.1093/neuonc/not072 ]
- Wienert S, Heim D, Kotani M, Lindequist B, Stenzinger A, Ishii M, Hufnagl P, Beil M, Dietel M, Denkert C, Klauschen F: CognitionMaster: an object-based image analysis framework. Diagn Pathol. 2013, 8: 34-10.1186/1746-1596-8-34.PubMed CentralView ArticlePubMed
- Gunduz C, Yener B, Gultekin SH: The cell graphs of cancer. Bioinformatics. 2004, 20 (Suppl 1): i145-i151. 10.1093/bioinformatics/bth933.View ArticlePubMed
- Gunduz C, Gultekin SH, Yener B: Augmented cell-graphs for automated cancer diagnosis. Bioinformatics. 2005, 21 (Suppl 2): ii7-ii12. 10.1093/bioinformatics/bti1100.
- West NP, Dattani M, McShane P, Hutchins G, Grabsch J, Mueller W, Treanor D, Quirke P, Grabsch H: The proportion of tumour cells is an independent predictor for survival in colorectal cancer patients. Br J Cancer. 2010, 102: 1519-1523. 10.1038/sj.bjc.6605674.PubMed CentralView ArticlePubMed
- Chang H, Han J, Borowsky A, Loss L, Gray JW, Spellman PT, Parvin B: Invariant delineation of nuclear architecture in Glioblastoma multiforme for clinical and molecular association. IEEE Trans Med Imag. 2013, 32 (4): 670-682.View Article
- Kayser K, Radziszowski D, Bzdyl P, Sommer R, Kayser G: Towards an automated virtual slide screening: theoretical considerations and practical experiences of automated tissue-based virtual diagnosis to be implemented in the internet. Diagn Pathol. 2006, 1: 10-10.1186/1746-1596-1-10.PubMed CentralView ArticlePubMed
- PLGA Foundation: Meta analysis low grade glioma database project. 2012, [http://www.fightplga.org/research/PLGA-Sponsored_Projects/MetaAnalysis]
- García Rojo M, Bueno G, Slodkowska J: Review of imaging solutions for integrated quantitative immunohistochemistry in the Pathology daily practice. Folia Histochem Cytobiol. 2009, 47 (3): 349-354.
- Rasband WS: ImageJ. 1997-2012. [http://imagej.nih.gov/ij/]
- ImageMagick Studio LLC: ImageMagick. 2013, [http://www.imagemagick.org/]
- GraphicsMagick Group: GraphicsMagick. 2013, [http://www.graphicsmagick.org/]
- Kong J, Cooper LAD, Wang F, Chisolm C, Moreno CS, Kurc TM, Widener PM, Brat DJ, Saltz JH: A comprehensive framework for classification of nuclei in digital microscopy imaging: An application to diffuse gliomas. Biomedical Imaging: From Nano to Macro, 2011 IEEE International Symposium on. 2011, 2128-2131.View Article
- Kayser K, Görtler J, Borkenfeld S, Kayser G: Grid computing in image analysis. Diagn Pathol. 2011, 6 (Suppl 1): S12-10.1186/1746-1596-6-S1-S12.PubMed CentralView ArticlePubMed
- Granier A, Olivier M, Laborie S, Vaudescal S, Baecker V, Tran-Aupiais C: WIDE (Web Images and Data Environment). 2013, [http://www.mri.cnrs.fr/index.php?m=81]
- Kayser K: Introduction of virtual microscopy in routine surgical pathology — a hypothesis and personal view from Europe. Diagn Pathol. 2012, 7: 48-10.1186/1746-1596-7-48.PubMed CentralView ArticlePubMed
- Goode A, Satyanarayanan M: A vendor-neutral library and viewer for whole-slide images. 2008, Tech. Rep. Technical Report CMU-CS-08-136, Computer Science Department, Carnegie Mellon University [http://reports-archive.adm.cs.cmu.edu/anon/2008/CMU-CS-08-136.pdf],
- Linkert M, Rueden CT, Allan C, Burel JM, Moore W, Patterson A, Loranger B, Moore J, Neves C, MacDonald D, Tarkowska A, Sticco C, Hill E, Rossner M, Eliceiri KW, Swedlow JR: Metadata matters: access to image data in the real world. J Cell Biol. 2010, 198 (5): 777-782.View Article
- Khushi M, Edwards G, de Marcos DA, Carpenter JE, Graham JD, Clarke CL: Open source tools for management and archiving of digital microscopy data to allow integration with patient pathology and treatment information. Diagn Pathol. 2013, 8: 22-10.1186/1746-1596-8-22.PubMed CentralView ArticlePubMed
- Sam Leffler S, the authors of LibTIFF: LibTIFF – TIFF Library and Utilities. 2012, [http://www.remotesensing.org/libtiff/]
- Lane TG, Vollbeding G: The Independent JPEG Group’s JPEG software. 2013, [http://www.ijg.org/]
- Lane TG, Vollbeding G, the authors of the libjpeg-turbo software: libjpeg-turbo. 2012, [http://libjpeg-turbo.virtualgl.org/]
- Schneider CA, Rasband WS, Eliceiri KW: NIH Image to ImageJ 25 years of image analysis. Nat Methods. 2012, 9: 671-675. 10.1038/nmeth.2089.View ArticlePubMed
- Sacha J: Image IO Plugin Bundle. 2004, [http://ij-plugins.sourceforge.net/plugins/imageio/]
- Sun Microsystems Inc: Java Advanced Library 1.1.3. 2006, [http://www.oracle.com/technetwork/java/current-142188.html]
- BigTIFF Design. 2012, [http://www.remotesensing.org/libtiff/bigtiffdesign.html]
- The BigTIFF File Format Proposal. 2008, [http://www.awaresystems.be/imaging/tiff/bigtiff.html]
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.