Pipeline Reference

Reference with all the possible options for the pipeline configuration file (“config.yml”), which uses the yaml format.

# global options must be defined

global:
  # Required: Path to a Fasta file
  sequences_file: /path/to/sequences.fasta
  # Required: String for output
  prefix: my_embeddings

  ## Optional: file manager
  # file_manager: [*filesystem]

  ## Optional: remap index simple (not via md5, this is not encouraged)
  # simple_remapping: [True, *False]

  ## Stages are executed in sequential order as they are outlined in this file.
  ## Stage names must be different!! If not: they will overwrite each_other
  ## The same stage type (e.g. embed,..) can be executed multiple types.
  ## Dependencies for a stage are defined in the dependencies parameter

  ## This config file includes options for initializing classes and options specific to the protocol

  ## Options notation:
  ## *: denotes the default option
  ## @: denotes that the file or directory will be downloaded and stored locally if not provided

stage_1:
  type: embed
  # Required: which embedder to use
  # Options: seqvec, prottrans_albert_bfd, prottrans_bert_bfd, prottrans_t5_bfd, prottrans_t5_uniref50,
  # prottrans_t5_xl_u50, prottrans_xlnet_uniref100, cpcprot, esm, esm1b, esm1v, plus_rnn, unirep, bepler
  protocol: seqvec

  # Optional: reduce embeddings to fixed size, per-protein. Comment out if not needed.
  # Note that you can always compute the reduced embeddings from the full embeddings
  # but some further stages (e.g. unsupervised extract) need this option
  reduce: True

  # Optional: discard per amino acid embeddings.
  # Setting this parameter to True will disable storing full size embeddings (per amino acid).
  # This parameter only works in combination with `reduce: True` or `embeddings_transformer_function`.
  # discard_per_amino_acid_embeddings: [True, *False]

  # Optional/Advanced: apply a transformation on the per-amino-acid embeddings
  # === This is an advanced parameter ===
  # This parameter will be "eval"-uated. It must be a callable. Most likely, you'll want to define lambda functions.
  # The input of the function is a per-amino-acid embedding (aka. an np array of shape n_layers*embedding_dimension*sequence_length).
  # You can use numpy functions via "np".
  # The result of the transformation will be stored in the transformed_embeddings_file.
  # The pipeline *WILL NOT* check file size of the prospective transformed_embeddings_file.
  # This parameter can be used in conjunction with `discard_per_amino_acid_embeddings`.
  # Two examples:
  #   - "lambda x: x[0].mean(0)" --> for SeqVec, this will return the mean pooled embedding of the first layer
  #   - "lambda x: x.max(0)" --> for ProtTrans-BERT-BFD, this will return the max pooled embedding, instead of mean pooled
  # embeddings_transformer_function:

  ## Mandatory for esm1v: This defines which of the five models of the ensemble will be used
  # ensemble_id: [1,2,3,4,5]

  #### Optional parameters to instantiate classes

  ### Optional for protocol: seqvec
  # weights_file: @/path/to/file
  # options_file: @/path/to/file
  ## The following parameter sets an upper bound on total AA to include when embedding many sequences.
  ## Adjust this parameter if CUDA runs out of memory! The default (15.000) works for a 1080 with 8GB RAM.
  # max_amino_acids: [*15000]
  ## The following parameters sets the amount of AA to include in a batch before writing to disk.

  ### Optional for protocol: fasttext, word2vec, glove, esm
  # model_file: @/path/to/file

  ### Optional for protocol: albert, bert, xlnet
  # model_directory: @/path/to/directory

  ### Optional for protocol: seqvec, prottrans_albert_bfd, prottrans_bert_bfd, prottrans_xlnet_uniref100, esm
  ## Set the following parameter to calculate embeddings on a specific device or a specific GPU on multi-GPU hosts.
  ## The default, "cuda", runs on the default GPU. You can switch to CPU with "cpu" or select a GPU on
  ## multi-GPU systems with "cuda:0" or "cuda:1" etc.
  ## See https://pytorch.org/docs/stable/tensor_attributes.html?highlight=device#torch.torch.device
  ## for a complete list
  ## unirep will use the cpu by default; Follow https://github.com/google/jax#pip-installation
  ## to run it on the gpu
  # device: cuda

  ## Save numbers with lower precision, so that they take only about half the
  ## storage space (replace float32 with float16).
  ## This makes predictions based on the embeddings less exact.
  # half_precision: [true, *false]

  ## For prottrans_t5_bfd, prottrans_t5_uniref50 and prottrans_t5_xl_u50 only:
  # Use the model in half precision mode (float16)
  ## We recommend activating this with T5, since tested GPU (Quadro RTX 3000) reduces memory consumption
  ## from 12GB to 7GB while the effect in benchmarks is negligible (±0.1 percentages points in different sets,
  ## generally below standard error)
  # half_precision_model: [true, *false]

stage_2:
  type: project
  # Required: which projection algorithm to use
  # Options: tsne, umap, pb_tucker
  protocol: tsne
  # Either depend on an embedding stage with reduced embeddings
  depends_on: stage_1
  # or define mapping and reduced embedding file:
  # reduced_embeddings_file: path/to/reduced_embeddings_file.h5
  # mapping_file: path/to/mapping_file.csv


  ### Optional for protocol: tsne
  # n_iter: *15000
  # perplexity: *6
  # n_jobs: *-1

  ### Optional for protocol: umap
  # min_dist: *0.6
  # spread: *1
  # n_neighbors: *15

  ### Optional for protocol: tsne and umap
  # metric: *'cosine'
  # n_components: *3
  # random_state: *420
  # verbose: *1

  ## Optional with pb_tucker: Postprocess reduced embeddings with tucker, which is a contrastive
  ## learning model trained to distinguish CATH superfamilies. It reduces the dimensionality from 1024 to 128.
  # model_file: @/path/to/file

stage_3:
  type: mutagenesis
  ## Required: Which language model to use
  ## Options: protbert_bfd_mutagenesis
  protocol: protbert_bfd_mutagenesis
  ## Optional: temperature for softmax, see https://arxiv.org/abs/1503.02531
  # temperature: [1*]

  ## Optional: Since we're running ProtBert, the common ProtBert options explained in `extract` are supported
  # model_directory: @/path/to/directory
  # device: cuda
  # half_precision: [true, *false]
  # half_precision_model: [true, *false]

stage_4:
  type: visualize
  ## Required: which graph to render
  ## Options: plotly, plot_mutagenesis
  protocol: plotly
  ## For plotly: Either depend on a project stage with projected embeddings file
  depends_on: stage_2
  ## or define projected_reduced_embeddings_file:
  # projected_reduced_embeddings_file: path/to/projected_reduced_embeddings_file.h5
  ## For plot_mutagenesis: Either depend on a mutagenesis stage
  # depends_on: stage_3
  ## or define residue_probabilities_file:
  # residue_probabilities_file: path/to/residue_probabilities_file.csv

  ## Optional for plotly: csv file with annotations
  ## csv must have header: identifier, label
  # annotation_file: path/to/annotation_file.csv

  ## Optional for plotly: hide proteins for which there is no annotation in the annotation file (only relevant if annotation file is provided)
  # display_unknown: [False, *True]

  ## Optional for plotly: set to True if in annotation_file identifiers correspond to sequence MD5 hashes
  ## if set to False (default), mapping will be performed on original identifiers.
  ## Where missing or duplicate, will be ignored
  # merge_via_index: [True, *False]

  ## Optional for plotly: 2D vs 3D plot
  # n_components: [2,*3]

stage_5:
  type: extract
  ## Required: which method to use.
  ## Current options:
  ##   - seqvec_from_publication (Uses models evaluated in https://doi.org/10.1186/s12859-019-3220-8 )
  ##   - bert_from_publication (Uses models evaluated in https://doi.org/10.1101/2020.07.12.199554 )
  ##   - unsupervised (Uses concepts presented in https://github.com/Rostlab/goPredSim )
  protocol: seqvec_from_publication

  ## The supervised extract (bert_from_publication and seqvec_from_publication) needs
  ## the full embeddings, i.e. `discard_per_amino_acid_embeddings` must be false in the
  ## embed stage, which is the default. For the unsupervised extract, you need the
  ## reduced embeddings, i.e. set `reduce: True`.
  ## Instead of an embed stage, you can also provide a file manually, either supply for unsupervised:
  # reduced_embeddings_file:
  ## or for bert_from_publication and seqvec_from_publication
  # embeddings_file:
  depends_on: stage_2

  ## Optional for protocol: seqvec_from_publication, bert_from_publication,
  ## will be downloaded if not supplied
  # secondary_structure_checkpoint_file: path/to/checkpoint.pt
  # subcellular_location_checkpoint_file: path/to/checkpoint.pt

  ## Required for protocol: unsupervised
  # reference_embeddings_file: path/to/embeddings.hd5
  # reference_annotations_file: path/to/annotation_file.csv

  ## Optional for protocol: unsupervised
  ## The following two options refer to the pairwise_distance function of scikit-learn 0.23.2
  ## https://scikit-learn.org/stable/modules/generated/sklearn.metrics.pairwise_distances.html#sklearn.metrics.pairwise_distances
  # n_jobs: [1*]
  # metric: [euclidean*]
  ## The following will define how many neighbours to consider to transfer annotations.
  ## k = 1 means transfer the annotations from the nearest neighbor
  ## k > 1 will result in merging the annotations of all k > 1 neighbors onto the target embedding
  # k_nearest_neighbours: [1*]
  ## The following informs the pipeline whether you want to keep the pairwise distance matrix file.
  ## This is a CSV containing pairwise distances between query and reference embeddings using your metric.
  ## The file can become quite big (e.g. UniProt Human (query) vs. SwissProt (reference) results in 45GB)
  ## By default, this file will be discarded, but you can decide to keep it to perform other calculations.
  # keep_pairwise_distances_matrix_file: [*False]