Skip to content

Efficient phylogenetic tree inference for massive taxonomic datasets

License

Notifications You must be signed in to change notification settings

citiususc/veryfasttree

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VeryFastTree

VeryFastTree is a new tool designed for efficient phylogenetic tree inference, specifically tailored to handle massive taxonomic datasets. It is a highly-tuned implementation based on the FastTree-2 tool that takes advantage of parallelization and vectorization strategies to speed up the inference of phylogenies for huge alignments.

Regarding the performance, for example, VeryFastTree (v4.0 - July 2023) is able to construct a tree on one server (two 32-core Intel Xeon Ice Lake 8352Y processors) using single precision arithmetic from an ultra-large one million taxa alignment in 36 hours, while our previous version (v3.0) and FastTree-2 require more than 5 days. That is, VeryFastTree-4.0 is more than 3x times faster than VeryFastTree-3.0 and FastTree-2, respectively.

It is important to highlight that VeryFastTree keeps unchanged the phases, methods and heuristics used by FastTree-2 to estimate the phylogenetic tree. In this way, it produces trees with the same topological accuracy than FastTree-2. In addition, unlike the parallel version of FastTree-2, VeryFastTree is deterministic.

To facilitate the adoption from the research community, VeryFastTree keeps exactly the same command line arguments than FastTree-2. In this way, it is only necessary to replace the call to FastTree-2 by a call to VeryFastTree using the same options to increase the overall performance.

VeryFastTree is now included as package in Bioconda, MacPorts and all Debian-based Linux distributions.

VeryFastTree has also Python bindings here.

If you use VeryFastTree, please cite the following works:

Release Notes:

  • v4.0 (July 2023):

    • Introduction of new thread levels for improved parallelization.
    • Enhanced performance through new parallel regions (e.g., ML Lengths, ML splits, LogLk, etc.).
    • Threads used in tree creation: Top hits, TopHitNJSearch, FastNJSearch, and ExhaustiveNJSearch (-slow).
    • Implementation of a faster tree partitioning approach with significant speed improvements.
    • Tree partitioning limited to NNI, SPR, and upProfiles computations for memory conservation:
    • Parallel tree traversal implemented for remaining parts.
    • Replacement of disk storage for profiles with Disk Computing.
    • Shared and reused Top upProfiles among threads for memory efficiency and accelerated sequential parts.
    • Improved non-deterministic mode with removal of mutex usage.
    • Optimized performance by parallelizing non-deterministic parts in deterministic mode:
    • Also implemented non-deterministic parts in deterministic mode for improved performance.
    • Deterministic mode now outperforms non-deterministic mode in terms of speed.
    • Tree partitioning method logging now hidden by default.
    • Support for Fastq format and libBZ2 compression.
    • Support for reading trees from NEXUS block trees.
    • Nvidia CUDA GPU computing support (experimental)
    • Introduced parallel compilation.
    • Incorporation of changes from FastTree-2.1.11.
    • Clang Support
    • Addressed critical errors and implemented substantial corrections.
  • v3.3.0 (merged into 4.0):

    • Deterministic mode now also parallelizes non-deterministic parts, but it requires more computation.
    • Tree partitioning algorithm is faster and has a partitioning cache.
  • v3.2.0 (December 2022):

    • Decrease in the peak memory usage.
    • Now profiles can be optionally stored on disk. It causes an important reduction in the memory usage.
    • All supported input files (NEXUS, Fasta and Phylip) supports Zlib compression.
  • v3.1.0 (December 2021):

    • NEXUS format is now supported.
    • Parallelization of SPR (Subtree-Prune-Regraft) moves.
    • New tree partitioning algorithm for NNI and SPR.
    • Choose between deterministic and non-deterministic parallelization (best performance).
    • Minor fixes.
  • v3.0 (May 2020):

    • AVX2 and AVX512 support.
    • NNI (Nearest-Neighbor-Interchange) parallelization.
    • Parallel computation of posterior distributions for each internal node.
    • Deterministic result.

Getting started

Requirements

All the libraries needed to compile VeryFastTree are included inside the lib folder. The other basic requirements are:

  • CMake v3.5+
  • C++11 compiler
    • GCC 5+ (GCC 4 is bugged)
    • Visual studio 2015 (previous versions with support for C++11 may work)
    • Clang (requires minimal support for C++11 and OpenMP)
  • make (Linux only)
  • CUDA Toolkit (CUDA only)

Configuring

CMake will generate the necessary configuration files for the compilation. In Linux/MacOS by default a native compilation is done for the considered architecture. This allows to detect automatically the hardware features. If the program is going to be executed on a different machine, this option must be disabled.

Options can be listed with cmake:

cmake -LH

// enable/disable system's processor architecture optimization (linux)
USE_NATIVE:BOOL=ON

// enable/disable SSE2 in Windows (Linux default)
USE_SEE2:BOOL=ON

// enable/disable SSE4.1
USE_SEE4:BOOL=OFF

// enable/disable AVX
USE_AVX:BOOL=OFF

// enable/disable AVX2
USE_AVX2:BOOL=OFF

// enable/disable AVX512
USE_AVX512:BOOL=OFF

// enable/disable CUDA
USE_CUDA:BOOL=OFF

// change CUDA Architecture
CUDA_ARCH:STRING=80

Example:

cmake -DUSE_NATIVE=OFF -DUSE_SEE4=ON . // Disable native compilation and use SSE 4.1

Building

Windows:

cmake [options] .
cmake --build

Linux/MacOS:

cmake [options] .
make
make install #Optional

Running VeryFastTree

To improve the usability and facilitate the adoption of VeryFastTree, it implements the same command interface than FastTree-2. It means that arguments have exactly the same behavior as in FastTree-2. All these arguments can be consulted with the -h option. As a consequence, to take advantage of the performance benefits provided by VeryFastTree is only necessary to replace the call to FastTree-2 by a call to VeryFastTree using the same options.

VeryFastTree accepts alignments in NEXUS, Fasta, Fastq or Phylip interleaved formats compressed with ZLib and libBZ2.

On the other hand, VeryFastTree has its own extra arguments which have been grouped in the Optimizations section. These arguments are related to the parametrization of the different parallelization and vectorization strategies included in VeryFastTree. Next we list and explain the new arguments available:

  • -threads [n] Number of threads (n) used in the parallel execution. If this option is not set, the corresponding value will be obtained from the environment variable OMP_NUM_THREADS. This is the same approach followed by FastTree-2. If n=1, VeryFastTree behaves in the same way than FastTree-2 compiled without the -DOPENMP flag.

  • -threads-level [level] Degree of parallelization:

    • If level is 0, VeryFastTree uses the same parallelization strategy as FastTree-2 with some new parallel blocks.
    • If level is 1, VeryFastTree uses parallel blocks that require additional memory for computation.
    • If level is 2, VeryFastTree accelerates the rounds of ML NNIs using its tree partitioning method.
    • If level is 3 (default), VeryFastTree performs more computations without preserving sequential order.
    • If level is 4, VeryFastTree accelerates the rounds of SPR steps using its tree partitioning method (it can only be used with datasets larger than 2^(sprlength + 2)).

    Note: Each level includes the previous ones, and computation at level 2 and above is performed in a different tree traverse order, so the result may change but is still correct.

  • -threads-mode [mode] Changes the mode of parallelization:

    • If mode is 0, VeryFastTree uses non-deterministic parts, some inspired by FastTree-2 but improved.
    • If mode is 1 (default), VeryFastTree only uses deterministic parallelization.

    Since version 4.0, deterministic algorithms are at least faster than non-deterministic ones, making deterministic the preferred choice.

  • -threads-ptw [n] (Partitioning Tendency Window) It sets the size of the partitioning tendency window used by the tree partitioning algorithm to determine when to stop searching. The window stores the last solutions and checks if a better solution can be found. Increasing the value allows the algorithm to explore the tree deeper and potentially find better solutions. The default value is 50.

  • -threads-verbose To show subtrees assigned to the threads and theoretical speedup, only with verbose > 0

  • -double-precision To use double precision arithmetic. Therefore, it is equivalent to compile FastTree-2 with -DUSE_DOUBLE.

  • -ext [type] It enables the vector extensions:

    • AUTO: (default) selects AVX2 when -double-precision is used and SEE3 otherwise. If one extension is not available, the previous level is used.
    • NONE: Operations are performed with the native programming language operators. In addition, loops are unrolled with the aim of providing hints to the compiler for applying some optimization (including vectorization).
    • SSE3: Arithmetic operations are performed using SSE3 vector intrinsics. Each instruction operates on 128 bit registers, which could contain four 32-bit floats or two 64-bit doubles.
    • AVX: Arithmetic operations are performed using AVX vector intrinsics. Each instruction operates on 256 bit registers, which could contain eight 32-bit floats or four 64-bits doubles.
    • AVX2: Similar to AVX, but some arithmetic operations are performed using additional AVX2 vector intrinsics not included in the AVX instruction set. Each instruction operates on 256 bit registers, which could contain eight 32-bit floats or four 64-bit doubles).
    • AVX512: Arithmetic operations are performed using AVX512 vector intrinsics. Each instruction operates on 512 bit registers, which could contain sixteen 32-bit floats or eight 64-bits doubles.
    • CUDA: (Experimental) Arithmetic operations are performed using NVIDIA CUDA.
  • -fastexp [implementation] This option is used to select an alternative implementation for the exponential function (ex), which has a significant impact on performance:

    • 0: (default) Use the exp function included in the built-in math library with double precision.
    • 1: Use the exp function included in the built-in math library with simple precision (not recommended together with -double-precision option).
    • 2: Use a very efficient and fast implementation to compute an accurate approximation of ex using double precision arithmetic.
    • 3: Use a very efficient and fast implementation to compute an accurate approximation of ex using simple precision arithmetic (not recommended together with -double-precision option).
  • -disk-computing If there is not enough available RAM to perform the computation, disk will be used to store extra data when it was not needed. Using disk to perform the computation will substantially increase the execution time.

  • -disk-computing-path [path] Like -disk-computing but using a custom path folder to store data.

  • -disk-dynamic-computing By default, disk computing only creates files associated with static data in RAM, which means that there is no significant impact on performance as long as there is available RAM. This option further reduces memory usage by storing dynamic data on disk. However, even if there is enough RAM, it will have a negative impact on performance due to the creation and deletion of files.