Inferring single cell trajectories with Monocle3 (R)
Author(s) | Julia Jakiela |
Editor(s) | Helena Rasche Pavankumar Videm Wendi Bacon |
Reviewers |
OverviewQuestions:Objectives:
How do I perform a trajectory analysis using Monocle3 in R?
What should I do when Galaxy’s Monocle tools are not enough?
How do I assign cell types to the clusters?
How do I infer lineage relationships between clusters, without a time series?
How do I perform batch correction and differential expression analysis in Monocle?
Requirements:
Identify which operations are necessary to transform an AnnData object into the files needed for Monocle
Describe the Monocle3 functions in R
Recognise steps that can be performed in R, but not with current Galaxy tools
Repeat the Monocle3 workflow and choose appropriate parameter values
Compare the outputs from Scanpy, Monocle in Galaxy and Monocle in R
Describe differential expression analysis methods
- Introduction to Galaxy Analyses
- tutorial Hands-on: Generating a single cell matrix using Alevin
- tutorial Hands-on: Combining single cell datasets after pre-processing
- tutorial Hands-on: Filter, plot and explore single-cell RNA-seq data with Scanpy
Time estimation: 3 hoursSupporting Materials:Published: Apr 12, 2023Last modification: Oct 28, 2024License: Tutorial Content is licensed under Creative Commons Attribution 4.0 International License. The GTN Framework is licensed under MITpurl PURL: https://gxy.io/GTN:T00336version Revision: 11
Introduction
This tutorial is the next one in the Single-cell RNA-seq: Case Study series. This tutorial focuses on trajectory analysis using monocle3, similar to the Monocle3 in Galaxy tutorial. However, in this tutorial we will use the R programming language that hides behind the user-friendly Galaxy tools. Sometimes you might encounter limitations when working with Galaxy tools, or you might want to make a wee modification that has to be done manually. It is therefore useful to be able to switch to R. If you do not feel confident using R, this tutorial is a good place to start. However, our tutorial is quite straightforward to follow and at the end you will feel like a programmer! On the other hand, if you are not confident with the biological or statistical theory behind trajectory analysis, check out the slide deck. With those resources (including the previous case study tutorials) you are well-equipped to go through this tutorial with ease. Let’s get started!
CommentThis tutorial is significantly based on the Monocle3 documentation.
Optional: Get data into a Galaxy history
In the Monocle3 in Galaxy tutorial, we showed that Monocle3 works great with annotated data, but what if your data is not annotated yet? Is it still possible to use Monocle? The answer is yes, Monocle can annotate cells according to their type.
First, we need to retrieve the appropriate data. We will continue to work on the case study data from a mouse model of fetal growth restriction Bacon et al. 2018 (see the study in Single Cell Expression Atlas and the project submission). We will use the filtered AnnData object, before normalisation and annotation, generated in the filtering tutorial as step 20: Filtered Object
. However for this tutorial, you don’t have to download any files on your computer or even import files to Galaxy! We will show you the whole analysis in R, starting from AnnData object. However, if you’d like to examine the datasets in a history, the instructions are below.
Hands-on: Optional data upload into Galaxy historyYou have three options for importing the input data into a Galaxy history.
You can import a history from: input history; Import the files from Zenodo; or Import the files from the shared data library (
GTN - Material
->single-cell
->Inferring single cell trajectories with Monocle3 (R)
):
- Open the link to the shared history
- Click on the Import this history button on the top left
- Enter a title for the new history
- Click on Copy History
https://zenodo.org/record/7455590/files/AnnData_filtered.h5ad
As an alternative to uploading the data from a URL or your computer, the files may also have been made available from a shared data library:
- Go into Data (top panel) then Data libraries
- Navigate to the correct folder as indicated by your instructor.
- On most Galaxies tutorial data will be provided in a folder named GTN - Material –> Topic Name -> Tutorial Name.
- Select the desired files
- Click on Add to History galaxy-dropdown near the top and select as Datasets from the dropdown menu
In the pop-up window, choose
- “Select history”: the history you want to import the data to (or create a new one)
- Click on Import
- Copy the link location
Click galaxy-upload Upload Data at the top of the tool panel
- Select galaxy-wf-edit Paste/Fetch Data
Paste the link(s) into the text field
Press Start
- Close the window
Check that the datatype is
h5ad
- Click on the galaxy-pencil pencil icon for the dataset to edit its attributes
- In the central panel, click galaxy-chart-select-data Datatypes tab on the top
- In the galaxy-chart-select-data Assign Datatype, select
h5ad
from “New type” dropdown
- Tip: you can start typing the datatype into the field to filter the dropdown menu
- Click the Save button
Rename galaxy-pencil the history to your name of choice.
Launching JupyterLab
warning Please note: this is only currently available on the usegalaxy.eu site.
Warning: Data uploads & JupyterLabThere are a few ways of importing and uploading data into JupyterLab. You might find yourself accidentally doing this differently than the tutorial, and that’s ok. There are a few key steps where you will call files from a location - if these don’t work from you, check that the file location is correct and change accordingly!
JupyterLab is a bit like RStudio but for other coding languages. What, you’ve never heard of RStudio? Then don’t worry, just follow the instructions! Saying that, for those that are confident with the R programming language and want to run this locally in the RStudio programming environment, we have included some extra details throughout that helped us run the code locally in RStudio. This tutorial is built to run in JupyterLab, however.
Hands-on: Launch JupyterLabCurrently JupyterLab in Galaxy is available on Live.useGalaxy.eu, usegalaxy.org and usegalaxy.eu.
Hands-on: Run JupyterLab
- Interactive Jupyter Notebook. Note that on some Galaxies this is called Interactive JupyTool and notebook:
- Click Run Tool
The tool will start running and will stay running permanently
This may take a moment, but once the
Executed notebook
in your history is orange, you are up and running!- Click on the User menu at the top and go to Active Interactive Tools and locate the JupyterLab instance you started.
- Click on your JupyterLab instance
If JupyterLab is not available on the Galaxy instance:
- Start Try JupyterLab
Welcome to JupyterLab!
Warning: Danger: You can lose data!Do NOT delete or close this notebook dataset in your history. YOU WILL LOSE IT!
Open the notebook
You have two options for how to proceed with this JupyterLab tutorial - you can run the tutorial from a pre-populated notebook, or you can copy and paste the code for each step into a fresh notebook and run it. The initial instructions for both options are below.
Hands-on: Option 1: Open the notebook directly in JupyterLab
- Open a
Terminal
in JupyterLab with File -> New -> Terminal
- Run
wget https://training.galaxyproject.org/training-material/topics/single-cell/tutorials/scrna-case_monocle3-rstudio/single-cell-scrna-case_monocle3-rstudio.ipynb
- Select the notebook that appears in the list of files on the left.
Remember that you can also download this notebook Jupyter Notebook from the galaxy_instance Supporting Materials in the Overview box at the beginning of this tutorial.
Hands-on: Option 2: Creating a notebook
- Select the R icon under Notebook
Save your file (File: Save, or click the galaxy-save Save icon at the top left)
If you right click on the file in the folder window at the left, you can rename your file
whateveryoulike.ipynb
Warning: You should Save frequently!This is both for good practice and to protect you in case you accidentally close the browser. Your environment will still run, so it will contain the last saved notebook you have. You might eventually stop your environment after this tutorial, but ONLY once you have saved and exported your notebook (more on that at the end!) Note that you can have multiple notebooks going at the same time within this JupyterLab, so if you do, you will need to save and export each individual notebook. You can also download them at any time.
Let’s crack on!
Installation
If you followed the tip tip above, you should already have your JupyterLab instance open. Before we start working on the tutorial notebook, we need to install required packages.
Hands-on: Installing the packages
- Navigate back to the
Terminal
- In the Terminal tab open, write the following, one line at a time:
conda install -y -c conda-forge -c bioconda r-monocle3
conda install -y -c conda-forge -c bioconda anndata
conda install -y -c conda-forge r-viridislite
conda install -y -c conda-forge bioconductor-biomart
- If you are asked at any point
Proceed ([y]/n)?
, typey
- surely we want to proceed!
Installation will take a while, so in the meantime, when it’s running, you can open the notebook and follow the rest of this tutorial there!
Monocle 3 runs in the R statistical computing environment. You will need R version 4.1.0 or higher, Bioconductor version 3.14, and monocle3 1.2.7 or higher to have access to the latest features. Here is the original code that you should run to install BiocManager and monocle3.
# Install Bioconductor and some of its dependencies
if (!requireNamespace("BiocManager", quietly = TRUE))
install.packages("BiocManager")
BiocManager::install(version = "3.14")
BiocManager::install(c('BiocGenerics', 'DelayedArray', 'DelayedMatrixStats',
'limma', 'lme4', 'S4Vectors', 'SingleCellExperiment',
'SummarizedExperiment', 'batchelor', 'HDF5Array',
'terra', 'ggrastr'))
# Install monocle3 through the cole-trapnell-lab GitHub:
install.packages("devtools")
devtools::install_github('cole-trapnell-lab/monocle3')
library(monocle3)Warning: Installation errorsIt may happen that you will encounter some problems with installation of monocle3 when using RStudio Galaxy instance or RStudio Cloud. It might be due to using older versions of R or required packages or lack of required dependencies. If it happens, you would need to carefully read the error messages and follow the suggestions. If you are facing any difficulties with installation process, it is recommended that you consult additional online resources. It is more likely that RStudio Cloud or Galaxy tool would fail rather than local RStudio. To make your analysis stress-free, you can follow the Jupyter Notebook instead, which should not give you installation issues.
Warning: Notebook-based tutorials can give different outputsThe nature of coding pulls the most recent tools to perform tasks. This can - and often does - change the outputs of an analysis. Be prepared, as you are unlikely to get outputs identical to a tutorial if you are running it in a programming environment like a Jupyter Notebook or R-Studio. That’s ok! The outputs should still be pretty close.
Setting up the environment and file upload
Once the installation is done, we should load the needed packages into our notebook. Navigate back to your notebook
. If you are using our prepopulated notebook, you can follow the tutorial from there. Otherwise, input the following into your fresh notebook.
install.packages("Rcpp") # needed for reduce_dimension to avoid AnnoyAngular error; library(Rcpp) might work as well depending on the version
library(monocle3)
library(biomaRt)
library(magrittr) # needed for %>%
library(viridisLite)
install.packages("anndata")
Alright, the libraries are here - so now let’s get our AnnData file.
Upload, view and modify the files
We can download the file into Jupyter notebook also using a Zendo link. The first argument is the download link and the second one is the name under which the downloaded file is saved:
download.file('https://zenodo.org/record/7455590/files/AnnData_filtered.h5ad', 'AnnData.h5ad')
Then we have to read in this h5ad file:
ann <- anndata::read_h5ad('AnnData.h5ad')
Now we store all the information we need in AnnData object. However, Monocle uses cell_data_set class to hold expression data, which requires three input files: expression_matrix
, cell_metadata
and gene_metadata
. Therefore, we have to somehow ‘transfer’ that information from our AnnData object to Monocle’s cell_data_set (cds). AnnData stores a data matrix X
together with annotations of observations obs
and variables var
, so we can extract those parameters and use them for further analysis.
expression_matrix <- ann$X
cell_metadata <- ann$obs
gene_metadata <- ann$var
If you already have files containing the expression matrix, genes and cells metadata, you can upload them to JupyLab and generate a cds file from them instead. For example, if you first downloaded the files from Galaxy, your files will have the
.tabular
extension. In this case, we will use theread.delim()
function to read them in. In this function, the first argument is the file path - in our case, the files are in the same folder as the notebook, so the file path is the same as the file name. You can always check that by right-clicking on the file and choosingCopy path
. The second argument,row.names=1
takes the column number of the data file from which to take the row names.# read in the files cell_metadata <- read.delim('cells.tabular', row.names=1) gene_metadata <- read.delim('genes.tabular', row.names=1) expression_matrix <- read.delim('expression.tabular', row.names=1)
QuestionWhy should we set
row.names=1
?This allows us to ensure that the expression value matrix has the same number of columns as the
cell_metadata
has rows and the same number of rows as thegene_metadata
has rows. Importantly, the row names of thecell_metadata
object should match the column names of the expression matrix and the row names of thegene_metadata
object should match the row names of the expression matrix.
If you are working in RStudio Cloud, you have to click on ‘Upload’ button in the right bottom window toolbar and choose already downloaded data files to upload. You should now see all three files in this window. You might want to rename the files to make their names shorter.
If you are using RStudio Galaxy tool, you can get data directly from your history by running:
# get the files from Galaxy history file.copy(gx_get(2), "cell_metadata") file.copy(gx_get(3), "gene_metadata") file.copy(gx_get(4), "expression_matrix")
The number in the brackets corresponds to the dataset number in your history, so make sure you put the right number for the corresponding file. We can specify the name of the fetched files in the quotation marks, as shown above. All three files should appear in the Files tab window.
If you are using RStudio locally, then you don’t have to bother about uploading the files – just run
file.choose()
and choose the corresponding file which you want to get path to:# get the file path cells_path <- file.choose() genes_path <- file.choose() expression_path <- file.choose()
You should now see the new variables in the Environment tab window.
According to the Monocle3 documentation, the expression_matrix
should have genes as rows and cells as columns. Let’s check if that’s the case here.
head(expression_matrix,c(5, 5)) # preview the content of the file by calling its the first 5 rows by 5 columns
We can see that in our matrix rows are cells and genes are columns, so we have to transpose the matrix simply using function t()
. But before doing so, we will change its type from dataframe to matrix - this is Monocle’s requirement to generate cell_data_set afterwards.
expression_matrix <- as.matrix(expression_matrix) # change the type to matrix
expression_matrix <- t(expression_matrix) # transpose the matrix
Another condition we have to satisfy is that one of the columns of the gene_metadata
should be named “gene_short_name”, which represents the gene symbol for each gene. Some functions won’t work without that. Do we have such a column? Let’s check.
head(gene_metadata) # preview the top ten rows of the file to check the name of the column containing gene symbols
The second column indeed contains gene symbols, but is called “Symbol” instead of “gene_short_name”. That can be easily changed by a simple assignment, as long as we know the number of the column that we want to modify. In our case the gene symbols are stored in column 2. We can access the column names by colnames()
.
colnames(gene_metadata)[2] <- 'gene_short_name' # change the column name
colnames(gene_metadata) # see the changes
Generating CDS object
Now let’s store our files in one object – the cell_data_set
. This is the main class used by Monocle to hold single cell expression data. The class is derived from the Bioconductor SingleCellExperiment class. Similar to Python’s AnnData, the cell_data_set stores a data matrix together with annotations of observations and variables. There are three ways of creating CDS object in monocle:
- Using
new_cell_data_set()
function with three data frames as arguments (not their paths!): expression matrix (can also be a sparseMatrix), cell metadata and gene metadata - Using
load_cellranger_data()
function providing the path to the folder containing 10X Genomics Cell Ranger output files. This function takes an argumentumi_cutoff
that determines how many reads a cell must have to be included - Using
load_mm_data()
function providing the paths to matrix file and two metadata files (features and cell information).
In this tutorial we will use the first option:
# create CDS object
cds <- new_cell_data_set(expression_matrix, cell_metadata, gene_metadata)
We are now ready to process our data!
Since Monocle’s CDS object is analogous to Python’s AnnData, why don’t we use some kind of conversion between those two formats? There is indeed a package called
sceasy
that helps easy conversion of different single-cell data formats to each other. However, when we tested this conversion on our dataset and then used Monocle to plot the expression of genes, the plots were not correct – the expression was shown to be identical throughout the sample. For comparison, Seurat did well when plotting gene expression of the same converted object! Although conversion functions are very handy, you have to be aware that their output might be interpreted differently by certain packages. Therefore, to make sure that the analysis is reliable, we decided to generate CDS object directly using Monocle’s function.
Additional step: adding genes symbols based on their IDs
Warning: Additional stepThis step is not necessary for the dataset we are working on but some users might find it helpful when analysing their own data.
If you remember the very first tutorial in the case study, we started with gene IDs and added gene symbols based on the Ensembl GTF file.
But what if we didn’t have the genes symbols in our CDS object and wanted to add them now? Of course - it’s possible! We will also base this annotation on Ensembl - the genome database – with the use of the library BioMart. We will use the same archive as in the Alevin tutorial (Genome assembly GRCm38) to get the gene names. Please note that the updated version (GRCm39) is available, but some of the gene IDs are not in that EnsEMBL database. The code below is written in a way that it will work for the updated dataset too, but will produce ‘NA’ where the corresponding gene name couldn’t be found.
cds_extra <- cds # assign our CDS to a new object for the demonstration purpose
head(rownames(fData(cds_extra))) # preview of the gene IDs as rownames
# get relevant gene names
library("biomaRt") # load the BioMart library
ensembl.ids <- rownames(fData(cds_extra)) # fData() allows to access cds rowData table
mart <- useEnsembl(biomart = "ENSEMBL_MART_ENSEMBL") # connect to a specified BioMart database and dataset hosted by Ensembl
ensembl_m = useMart("ensembl", dataset="mmusculus_gene_ensembl", host='https://nov2020.archive.ensembl.org')
# The line above connects to a specified BioMart database and dataset within this database.
# In our case we choose the mus musculus database and to get the desired Genome assembly GRCm38,
# we specify the host with this archive. If you want to use the most recent version of the dataset, just run:
# ensembl_m = useMart("ensembl", dataset="mmusculus_gene_ensembl")
genes <- getBM(attributes=c('ensembl_gene_id','external_gene_name'),
filters = 'ensembl_gene_id',
values = ensembl.ids,
mart = ensembl_m)
# The line above retrieves the specified attributes from the connected BioMart database;
# 'ensembl_gene_id' are genes IDs,
# 'external_gene_name' are the genes symbols that we want to get for our values stored in ‘ensembl.ids’.
# see the resulting data
head(genes)
# replace IDs for gene names
gene_names <- ensembl.ids
count = 1
for (geneID in gene_names)
{
index <- which(genes==geneID) # finds an index of geneID in the genes object created by getBM()
if (length(index)==0) # condition in case if there is no corresponding gene name in the chosen dataset
{
gene_names[count] <- 'NA'
}
else
{
gene_names[count] <- genes$external_gene_name[index] # replaces gene ID by the corresponding gene name based on the found geneID’s index
}
count = count + 1 # increased count so that every element in gene_names is replaced
}
# store the gene names in our CDS object in a new column gene_short_name_extra
fData(cds_extra)$gene_short_name_extra <- gene_names
# see the changes
fData(cds_extra)
If you are working on your own data and it’s not mouse data, you can check available datasets for other species and just use relevant dataset in useMart()
function.
listDatasets(mart) # available datasets
Warning: Ensembl connection problemsSometimes you may encounter some connection issues with Ensembl. To improve performance Ensembl provides several mirrors of their site distributed around the globe. When you use the default settings for useEnsembl() your queries will be directed to your closest mirror geographically. In theory this should give you the best performance, however this is not always the case in practice. For example, if the nearest mirror is experiencing many queries from other users it may perform poorly for you. In such cases, the other mirrors should be chosen automatically.
Monocle workflow
Do you remember the Monocle workflow introduced in the previous tutorial? Here is a recap:
Pre-processing
Let’s start with normalisation and pre-processing that can be performed using the function preprocess_cds()
. The argument num_dim
is the number of principal components that will be used. You can check that you’re using enough PCs to capture most of the variation in gene expression across all the cells in the data set. Note that “PCA” is the default method of pre-processing in Monocle3, so although we can specify this in our function, we don’t have to.
Note that this step can take awhile - around 5 or so minutes due to the high number of PCs calculated. Feel free to use a num_dim
value around 100 to accelerate preprocessing and compare the results.
# PCA pre-processing with 210 principal components
cds_preprocessing <- preprocess_cds(cds, method = "PCA", num_dim = 210)
# plot the variation in gene expression vs PCA components
plot_pc_variance_explained(cds_preprocessing)
The plot shows that actually using more than ~100 PCs captures only a small amount of additional variation. However, if we look at how the cells are plotted on 2D graph when using different values of PCs, it is easier to visualise how the num_dim
actually affects the output. We will use the value of 210, which, compared to the results from the previous tutorial, makes the most sense for our dataset.
Batch correction and Dimensionality reduction
Our dataset actually comprises data from 7 samples, so there is a risk that batch effects will impact analysis. Batch effects are systematic differences in the transcriptome of cells measured in different experimental batches. However, we can use Monocle to deal with that!
First, let’s check how our dataset looks in terms of batch effects. We can do that by colouring the cells by batch. This information is stored in our CDS object from the cell_metadata
file. Before asking Monocle to plot anything, let’s check the exact column name of the batch information column.
colnames(colData(cds_preprocessing)) # check column names
In our case it’s indeed ‘batch’, but your data might have another name (eg. “plate”, etc.), so make sure you put the correct argument value.
In order to plot our cells on a 2D graph, we need to reduce the numbers of dimensions. The previous tutorial introduced the methods of dimensionality reduction in Monocle. You can replicate what we did in Galaxy to compare the output of dimensionality reduction using different methods, simply by changing the reduction_method
argument. Options currently supported by Monocle are “UMAP”, “tSNE”, “PCA”, “LSI”, and “Aligned”. However, as for now, let’s just recall that UMAP gave the best results, so we will use UMAP here as well.
# reduce dimension first
cds_preprocessing_UMAP <- reduce_dimension(cds_preprocessing, reduction_method = "UMAP", preprocess_method = "PCA")
# then produce a plot to check for batch effects
plot_cells(cds_preprocessing_UMAP, color_cells_by="batch", label_cell_groups=FALSE)
We can see that upper and lower right branches mostly consist of N705 and N706, so indeed batch correction might be helpful. Let’s run this.
# perform batch correction
cds_batch <- align_cds(cds_preprocessing_UMAP, preprocess_method = "PCA", alignment_group = "batch")
To see the changes, we have to run UMAP again, but this time on the aligned dataset. We will specify that preprocess_method
as “Aligned” and not “PCA”. Monocle would use the “Aligned” argument automatically if no preprocess_method
was specified.
# dimensionality reduction after alignment
cds_red_dim <- reduce_dimension(cds_batch, preprocess_method = "Aligned", reduction_method = "UMAP")
# see the batch correction effect on a plot
plot_cells(cds_red_dim, color_cells_by="batch", label_cell_groups=FALSE)
QuestionDoes your plot look the same as the one in the Figure?
Your plot might be slightly different to the one shown in the Figure but this is fine, as long as you see analogical patterns. Some libraries that you’re using might have been updated, giving non-identical output. However, the principle behind the analysis is still the same, so you can peacefully follow the tutorial. Just keep your eyes open and… THINK!
Do you see this? It’s amazing! Batch correction did a great job here! Now the dataset is nicely aligned, and the cells from all the samples are evenly spread throughout the whole dataset. It is worth mentioning that removing batch effects was done using mutual nearest neighbor alignment, a technique introduced by John Marioni’s lab (Haghverdi et al. 2018) and supported by Aaron Lun’s package batchelor. Also, due to the machine learning elements of the code of this technique - as well as the fact that packages are updated regularly - your plots may not look identical to the ones pictured here. Nevertheless, the interpretation should be the same - the batch corrected plot should show better batch distribution than the uncorrected one.
Now we can move to the next step and perform dimensionality reduction.
When creating graphs, we sometimes use labels and sometimes just a legend. You can choose whichever you think makes the data clear and readable.
If you want to use a legend, then specify an argument
label_cell_groups=FALSE
in the functionplot_cells()
.The labels are set automatically, but if you want to change their size (default labels are tiny), use the argument
group_label_size
.
Clustering: clusters
We want to get some information about cell types, don’t we? In order to do so, we have to cluster our cells first. Monocle uses a technique called “community detection” (Traag et al. 2019) to group cells. This approach was introduced by Levine et al. 2015 as part of the phenoGraph algorithm. Monocle also divides the cells into larger, more well separated groups called partitions, using a statistical test from Wolf et al. 2019, introduced as part of their PAGA algorithm.
Clusters are particularly useful while trying to assign cells to a certain type, because they are based on the similarity in gene expression. The relationships between different clusters are analysed to identify possible trajectories.
Partitions, meanwhile, are larger groups of cells that usually contain several clusters. Trajectory inference is performed only within one partition, so it is essential that all the cells that we want to analyse in pseudotime belong to the same partition.
Therefore, let’s perform clustering and visualise the resulting clusters.
# clustering
cds_auto_cluster <- cluster_cells(cds_red_dim, reduction_method = "UMAP")
# see the clusters
plot_cells(cds_auto_cluster, reduction_method = "UMAP", color_cells_by = 'cluster', group_label_size = 5)
When using standard igraph louvain clustering, the value of resolution parameter is by default set to NULL, which means that it is determined automatically. Although the resulting clusters are OK, it would be nice to get some more granularity to identify cell types more specifically. The higher the resolution value, the more clusters we get. We will set the resolution value to 0.0002, but you are very welcome to try different values to see the changes.
# clustering with changed resolution value
cds_clustered <- cluster_cells(cds_red_dim, reduction_method = "UMAP", resolution = 0.0002)
# see the new clusters
plot_cells(cds_clustered, reduction_method = "UMAP", color_cells_by = 'cluster', group_label_size = 5)
Clustering: partitions
OK, what about partitions? They were also created during the clustering step and it’s important to check them before calculating the trajectory because it is performed only within one partition. It is essential that all the cells that we want to analyse in pseudotime belong to the same partition.
# see the partitions
plot_cells(cds_clustered, reduction_method = "UMAP", color_cells_by = 'partition', label_cell_groups=FALSE)
While your plot might be slightly different due to package updates, we can see that there are 3 partitions identified in cds_clustered
object. Ideally, we would like to combine partitions 1 and 2 to draw a trajectory through all those cells (we can ignore cells in outlier partition). Sometimes using the default values might result in multiple partitions while you only need one. Then you would have to change the q-value cutoff in partition_qval
. The default is 0.05 and by increasing this value you can increase the span of partitions, meaning that you would get fewer partitions. When trying different values of q-value, you also have to check if the clusters didn’t change. It’s all about finding a balance between the value of resolution
and partition_qval
so that both clusters and partitions are satisfactory enough for downstream analysis. Let’s try that on our dataset.
# changing the partition q-value
cds_clustered <- cluster_cells(cds_red_dim, reduction_method = "UMAP", resolution = 0.0002, partition_qval = 1)
# see the partitions with the changed q-value
plot_cells(cds_clustered, reduction_method = "UMAP", color_cells_by = 'partition', label_cell_groups=FALSE)
# check if clusters didn't change
plot_cells(cds_clustered, reduction_method = "UMAP", color_cells_by = 'cluster', label_cell_groups=FALSE)
QuestionHave the clusters change after changing the partition q-value?
It might be the case that after changing partition q-value, you will notice that additional clusters appeared. In that situation, you might either play around the
resolution
andpartition_qval
values, go forward with the current clustering (adjusting the parameters accordingly), or check the other method of assigning cells to one partition given below.
Now we have cells of interest in one partition, we still have reasonable clusters, so now we can learn the trajectory. However, in some cases even this method might not be enough. Then, there is a last resort… assigning cells to a partition manually.
Additional step: assigning cells to one partition
Warning: Additional stepThis step is not necessary for the dataset we are working on but some users might find it helpful when analysing their own data.
Let’s assume we have 4 partitions that cannot be extended to one big partition using partition_qval
(it might sound unreasonable but it does happen!) and we are desperate to have all the cells in one partition to draw a trajectory through all of them. We can simulate the initial dataset by setting partition_qval
value to 0.0001.
# simulate the dataset
cds_partitions_extra <- cluster_cells(cds_red_dim, reduction_method = "UMAP", partition_qval = 0.0001)
# see the simulated partitions
plot_cells(cds_partitions_extra, reduction_method = "UMAP", color_cells_by = 'partition', label_cell_groups=FALSE)
# store ‘1’ the number of times equal to the number of cells
big_partition <- c(rep(1,length(cds_partitions_extra@colData@rownames)))
# take the barcode of each cell and assign ‘1’ to each of them (now the ‘ones’ are named)
names(big_partition) <- cds_partitions_extra@colData@rownames
# convert from numeric to factor
big_partition <- as.factor(big_partition)
# assign the created barcodes-partition list to the location where information about partitions is stored in CDS object
cds_partitions_extra@clusters$UMAP$partitions <- big_partition
# see the new partition assignment
plot_cells(cds_partitions_extra, reduction_method = "UMAP", color_cells_by = 'partition', label_cell_groups=FALSE)
As you saw above, we used
@
and$
operators to navigate in CDS object. What is the difference between them?$
operator is used to access one variable/column, while@
extracts the contents of a slot in an object with a formal (S4) class structure. If you useView()
function on our CDS object, you will see the elements of type “S4” that you can access using@
operator. The suggestion list of elements that can be accessed should also pop up as you type the name of the object followed by the corresponding operator.
Assigning cell types
There are two main approaches to assigning cell types to clusters that we’ve just identified – supervised and unsupervised, both based on gene expression in each cluster.
Supervised approach
The supervised approach relies on prior knowledge of which cell types to expect. We can simply check the expression of marker genes specific to the expected cell types. Let’s then check the markers mentioned in the original paper Bacon et al. 2018.
Marker | Cell type |
---|---|
Il2ra | Double negative (early T-cell) |
Cd8b1, Cd8a, Cd4 | Double positive (middle T-cell) |
Cd8b1, Cd8a, Cd4 - high | Double positive (late middle T-cell) |
Itm2a | Mature T-cell |
Aif1 | Macrophages |
Hba-a1 | RBC |
To plot the expression of all those genes across our dataset in one go, we will pass a vector with the names of those markers into a parameter genes
in the plot_cells
function.
# expression of marker genes across the sample
plot_cells(cds_clustered, genes=c('Il2ra','Cd8b1','Cd8a','Cd4','Itm2a','Aif1','Hba-a1'), reduction_method = "UMAP")
Question: Genes, cell types and clustersBased on the gene expression graph that we just generated, the table above and your knowledge from the previous tutorials, how would you assign the clusters?
Keep in mind that these results refer to our numbered clusters, while yours might be slightly different.
Il2ra
(DN): mostly expressed in cluster 4Cd8b1, Cd8a
(DP middle): expressed in clusters 1, 6, and highly in cluster 2Cd4
(DP late): average expression in clusters 1, 6, 2 and high expression in cluster 5Itm2a
(T-mat): expressed in cluster 3Aif1
(macrophages): barely anything here, minimal expression spread across the sample with some more cells in cluster 4 and 3 – not enough to form a distinct cluster though). In theory, we shouldn’t have any macrophages in our sample. If you remember from the previous tutorials, we actually filtered out macrophages from the sample during the processing step, because we worked on annotated data. When analysing unannotated data, we could only assign macrophages and then filter them out, provided that Monocle clusters them into a separate group. As you can see, it’s not the case here, so we will just carry on with the analysis, interpreting this as a contamination.Hba-a1
(RBC): appears throughout the entire sample in low numbers suggesting some background contamination of red blood cell debris in the cell samples during library generation, but also shows higher expression in a distinct tiny bit of cluster 3, at the border between clusters 1 and 5. However, it’s too small to be clustered into a separate group and filtered out in this case. If you remember, this gene was found to be expressed in the previous Scanpy tutorial also in low numbers across the sample, and in the other Monocle tutorial (using Galaxy tools and annotated data) algorithms allowed us to gather the cells expressing that gene into a distinct group. Our result now sits somewhere in between.
Having identified which cluster corresponds to a specific cell type, we can finally run some code to add the annotation to our CDS object. First, we will create a new column called cell_type
in colData()
- this is where the information about the cells is stored (eg. batch, genotype, sex, etc) - and initialize it with the values of clusters. Then, we will get the dplyr
package which will be used for cluster annotation.
# just to keep the objects tidy and not overwrite them so that you can come back to any point of the analysis
cds_annotated <- cds_clustered
# create a new column ‘cell_type’ and initialise it with clusters values
colData(cds_annotated)$cell_type <- as.character(clusters(cds_annotated))
# annotate clusters
colData(cds_annotated)$cell_type <- dplyr::recode(colData(cds_annotated)$cell_type,
'1'='DP-M1', # double positive – middle T-cell (1st cluster)
'2'='DP-M2', # double positive – middle T-cell (2nd cluster)
'3'='T-mat', # mature T-cell
'4'='DN', # double negative – early T-cell
'5'='DP-L', # double positive – late middle T-cell
'6'='DP-M3', # double positive – middle T-cell (3rd cluster)
'7'='Unknown') # no info for now, so call it ‘Unknown’
'8'='Unknown') # no info for now, so call it ‘Unknown’
# check the annotation
plot_cells(cds_annotated, color_cells_by="cell_type", label_cell_groups=FALSE)
Unsupervised approach
But what if we don’t have any reference that we can use to assign our clusters? In that case, we will turn to the mentioned unsupervised approach - we will check what are the specifically expressed genes for each cluster. Then we can identify the cell types by looking up what cell types contain those genes. That’s a more tedious process, but sometimes can lead to exciting and unexpected results.
We will use Monocle’s function top_markers()
and store the information about specifically expressed genes for each cluster in the data frame marker_test
.
# find top marker genes in each cluster
marker_test <- top_markers(cds_clustered, group_cells_by="cluster", reduction_method = "UMAP", reference_cells=1000, cores=8)
You can group the cells by any categorical variable in colData(cds_clustered)
. The parameter reference_cells
is used to accelerate the marker significance test at some cost in sensitivity. It works by randomly selecting a specified number of cells and performing the marker significance test against a chosen set of cells. If your dataset is not massively big, you might skip this parameter as it wouldn’t help much.
QuestionWhat are the variables stored in
marker_test
data frame and what do they mean?Those are 10 variables that you can check easily either using colnames(marker_test) or
View(marker_test)
which will also display all the corresponding values.
gene_id
- Ensembl gene IDgene_short_name
- short name of the gene corresponding to its IDcell_group
- a group to which the cell belongs, specified in thegroup_cells_by
argumentmarker_score
- numeric vector of marker scores as the fraction expressing scaled by the specificity. The value ranges from 0 to 1mean_expression
- numeric vector of mean normalized expression of the gene in the cell groupfraction_expressing
- numeric vector of fraction of cells expressing the gene within the cell groupspecificity
- numeric vector of a measure of how specific the gene’s expression is to the cell group based on the Jensen-Shannon divergence. The value ranges from 0 to 1.pseudo_R2
- numeric vector of pseudo R-squared values, a measure of how well the gene expression model fits the categorical data relative to the null model. The value ranges from 0 to 1.marker_test_p_value
- numeric vector of likelihood ratio p-values; p-value is an area in the tail of a distribution that tells you the odds of a result happening by chancemarker_test_q_value
- numeric vector of likelihood ratio q-values; q-value is a p-value that has been adjusted for the False Discovery Rate (FDR)
We can now use data in marker_test
to rank the cells based on one of the specificity metrics and take the top gene(s) for each cluster. We will filter the expressing cells that constitute more than 10% of the cell group and we will take one gene in each cluster with the highest pseudo_R2
value (you can of course modify this value and choose more genes to be selected).
# filter the ‘marker_test’ data frame
top_specific_markers <- marker_test %>%
dplyr::filter(fraction_expressing >= 0.10) %>% # set the fraction of expressing cells
dplyr::group_by(cell_group) %>% # set a group to which the cells belong
dplyr::top_n(1, pseudo_R2) # set the number of top genes and the variable from 'marker_test' to rank by
# store the names of the marker genes
# you can also use IDs, the conversion to gene names should happen automatically when plotting
top_marker_names <- unique(top_specific_markers %>% dplyr::pull(gene_short_name))
Now we have all elements to plot the expression and fraction of cells that express found markers in each group.
plot_genes_by_group(cds_clustered, # our CDS object
top_marker_names, # a list of gene names to show in the plot
group_cells_by="cluster", # how to group cells when labeling
ordering_type="maximal_on_diag") # how to order the genes / groups on the dot plot
If you notice that on your dot plot any cluster has more genes than you specified (here we set one gene per cluster
top_n(1, pseudo_R2)
), go back to the code and create a new cell, right after assigningtop_specific_markers
object and just beforetop_marker_names
. In the new cell, paste and execute the following:top_specific_markers <- top_specific_markers %>% dplyr::distinct(pseudo_R2, .keep_all=TRUE) # select only one row if there are multiple rows with the same value in 'pseudo_R2' column
Then you can execute
top_marker_names
again and see how the plot looks like now. Better, right? This problem may arise when several genes in one cluster have the same values of specific variable from ‘marker_test’ (in our case we chosepseudo_R2
). It might likely happen in small and quite unsignificant clusters.
Look at this – we have identified some more marker genes specific to each cluster! However, sometimes it happens that the found genes are not as specific as one would expect, and they appear across the whole sample. Therefore, it is a good idea to plot all those marker genes and check how they appear in the bigger picture.
# plot the identified genes to check their expression
plot_cells(cds_clustered, genes=c('Pcdhgc4','Pcdhga5','Gm5559','Gm10359','Ccr9','Cd8b1','Plac8', 'Il2ra', 'Cd52', 'Tmsb10', 'Mki67', 'Hmgb2', 'Pclaf', 'Npm1'), reduction_method = "UMAP")
Further steps from now would include reviewing literature and checking what cell types correspond to the genes expressed in each cluster. Then you can annotate your clusters in the same way as shown above. Once you have your clusters annotated, Monocle can generate a file of marker genes for the identified cell types. This file can be then used with Garnett, a software toolkit for automatically annotating cells. We will not go through this in the current tutorial, but we will generate the file of marker genes.
# use ‘top_markers()’ again, now grouping cells by the assigned cell type
assigned_marker_test <- top_markers(cds_annotated,
group_cells_by="cell_type",
reference_cells=1000,
cores=8)
# filter these markers according to how stringent you want to be
garnett_markers <- assigned_marker_test %>%
dplyr::filter(marker_test_q_value < 0.05 & specificity >= 0.25) %>%
dplyr::group_by(cell_group) %>%
dplyr::top_n(5, marker_score)
# exclude genes that are good markers for more than one cell type:
garnett_markers_filtered <- garnett_markers %>%
dplyr::group_by(gene_short_name) %>%
dplyr::filter(dplyr::n() == 1)
# generate a file of marker genes
generate_garnett_marker_file(garnett_markers_filtered, file="./marker_file.txt")
A new file should appear in the ‘Files’ window. If you click on it, you will see the cell types and their corresponding marker genes, satisfying your chosen conditions.
Note that you can use the above block of code to generate file with the marker genes for unannotated CDS object to help you identify and check specifically expressed genes – you’d only have to change
group_cells_by
parameter from “cell_type” to “cluster”.
If you are working in RStudio locally, you might want to try a great function
choose_cells()
, which allows you to make a subset of the CDS object containing only cells of interest to investigate certain clusters more in-depth. It only works in interactive mode, so can be only used locally - when you call it, then a pop-up window will appear where you can choose cells to subset.# create a CDS subset cds_subset <- choose_cells(cds_clustered)
Now the chosen cluster is stored as a separate CDS object and you can analyse it independently, using the methods described above.
Trajectory inference
It’s time to perform trajectory analysis! First, let’s learn the trajectory graph. With an argument use_partition
we can specify if we want to learn a disjoint graph (value TRUE - default) in each partition, or to learn a single graph (value FALSE) across all partitions. The thing is, we can visualise the cells in pseudotime only if they belong to one partition. This is why it is important to make sure that all the cells that you want to analyse in pseudotime belong to one partition.
In one of the previous sections we’ve already prepared the partitions for trajectory inference - we assigned all the cells of interest into one partition by changing the q-value. As a result, we got two partitions – one containing only the cells that we labeled as ‘Unknown’ and another including all other cells. That is a perfect assignment – we’d like to focus only on maturation of T-cells, so we don’t want to connect those two partitions (therefore we specify use_partition=TRUE
to get a disjoint graph) and we will work downstream only on the graph going through annotated T-cells.
# learn trajectory graph
cds_trajectory <- learn_graph(cds_annotated, use_partition=TRUE)
# visualise the learned trajectory
plot_cells(cds_trajectory,
color_cells_by = "cell_type",
label_cell_groups=FALSE,
label_groups_by_cluster=FALSE,
label_leaves=FALSE,
label_branch_points=FALSE)
We have to tell Monocle where to start ordering the cells, ie. when we expect the analysed biological process to begin. Thanks to our biological knowledge, we know that the beginning of the trajectory should be at DN cluster. There are a couple of ways to specify the root cells:
Option 1: Root nodes
Here, you will use root_pr_nodes
argument in order_cells()
function.
To find the names of the principal points, you have to plot the learned trajectory again, specifying label_principal_points = TRUE
# specifying root cells: `root_pr_nodes` argument - check the principal points
plot_cells(cds_trajectory,
color_cells_by = "cell_type",
label_cell_groups=FALSE,
label_groups_by_cluster=FALSE,
label_leaves=FALSE,
label_branch_points=FALSE,
label_principal_points = TRUE, # set this to TRUE
graph_label_size=3)
You can see now the principal points and their labels in the form Y_number
. Pick the principal point in the cluster that you expect to be the beginning of the trajectory and type its name in the root_pr_nodes
argument when calling order_cells()
function.
# specifying root cells: `root_pr_nodes` argument - use the relevant principal point
cds_order_1 <- order_cells(cds_trajectory, root_pr_nodes='Y_14')
There is also a helper function to identify the root principal points based on the annotated cell types. This function uses pr_graph_cell_proj_closest_vertex
which is just a matrix with a single column that stores for each cell, the ID of the principal graph node it’s closest to.
# a helper function to identify the root principal points
get_correct_root_state <- function(cds, cell_phenotype, root_type){
cell_ids <- which(pData(cds)[, cell_phenotype] == root_type)
closest_vertex <- cds@principal_graph_aux[["UMAP"]]$pr_graph_cell_proj_closest_vertex
closest_vertex <- as.matrix(closest_vertex[colnames(cds), ])
root_pr_nodes <- igraph::V(principal_graph(cds)[["UMAP"]])$name[as.numeric(names
(which.max(table(closest_vertex[cell_ids,]))))]
root_pr_nodes
}
# call the function to automatically find the node in the principal graph where our DN cells reside
DN_node_id = get_correct_root_state(cds_trajectory, cell_phenotype = 'cell_type', "DN")
DN_node_id # check the node found
# order cells using the helper function output
cds_order_1_helper <- order_cells(cds_trajectory, root_pr_nodes = DN_node_id)
Option 2: Root cells
Here, you will use root_cells
argument in order_cells()
function.
Specify a vector of starting cell IDs. You can provide only one cell as well as all cells of a given type.
# find the names of all cells belonging to a certain type, identified as a beginning of a trajectory
starting_cell_type <- 'DN'
index <- which(cds_trajectory@colData$cell_type == starting_cell_type)
DN_cells <- colnames(cds_trajectory)[index]
# alternatively, if you work on unannotated data, you can use the number of the cluster that should be used as the beginning of the trajectory and pass it in the ‘root_cells’ argument
starting_cluster <- colnames(cds_trajectory[,clusters(cds_trajectory) == 4])
# order cells
cds_order_2 <- order_cells(cds_trajectory, root_cells = DN_cells)
If you are working in RStudio locally, you can use
order_cells()
function in the interactive mode. The pop-up window should appear and then you can simply select root nodes in the cluster with the cell type identified as the beginning of the trajectory.# specifying root cells: pop up window cds_order_1 <- order_cells(cds_trajectory)
Plotting in pseudotime
You can use any cds_order
object for the downstream analysis. Let’s pick one and assign it to an object with a shorter and more general name.
cds_order <- cds_order_1_helper
The function order_cells()
calculates pseudotime for each cell, so we can now visualise the T-cell maturation process in pseudotime! Additionally, we can access it and store in our CDS object for further analysis.
# plot cells in pseudotime
plot_cells(cds_order,
color_cells_by = "pseudotime",
label_cell_groups=FALSE,
label_leaves=FALSE,
label_branch_points=FALSE)
# access pseudotime calculated for each cell and store it alongside cell metadata
pseudotime <- pseudotime(cds_order)
cds_order@colData$pseudotime <- pseudotime
We can now see how our hard work has come together to give a final pseudotime trajectory analysis, which starts at double negative cells, then gently switches to double positives: from middle to late T-cells, and ends up on mature T-cells.
Differential expression analysis
There are two approaches for more advanced differential analysis in Monocle:
- Regression analysis: using
fit_models()
, you can evaluate whether each gene depends on variables such as time, treatments, etc. - Graph-autocorrelation analysis: using
graph_test()
, you can find genes that vary over a trajectory or between clusters.
In this section, for some examples, we will work on the subset of our CDS object because the computation time for the whole dataset would take quite some time. Let’s make the subset CDS containing the information about the genes listed in the table in the previous section, so instead having 15395 elements, it will have only 7.
# make the subset CDS
test_genes=c('Il2ra','Cd8b1','Cd8a','Cd4','Itm2a','Aif1','Hba-a1')
cds_subset <- cds_order[rowData(cds_order)$gene_short_name %in% test_genes,]
Some more plotting
Monocle also provides some easy ways to plot the expression of a small set of genes grouped by the factors you use during differential analysis.
For example plot_genes_violin()
allows us to create violin plots which are quite common in the field. Therefore let’s visualise how the gene expression changes between the cell types.
# produce violin plots
plot_genes_violin(cds_subset, group_cells_by="cell_type", ncol=2)
When analysing the above violin plots, we will realise that the results are consistent with the detailed analysis in the previous tutorials, as well as findings in current tutorial.
Another great function plot_genes_in_pseudotime()
takes a small set of genes and shows their dynamics as a function of pseudotime.
plot_genes_in_pseudotime(cds_subset, color_cells_by="cell_type", min_expr=0.5)
Regression analysis - advanced
We will use a function which fits a generalized linear model for each gene in a CDS object. We have to specify the model formula which is any term that exists as a column in colData()
. We want to test genes that differ between cell types and batches, so we will use “~cell_type + batch” argument. Then, we extract a table of coefficients from each model. coefficient_table()
tests whether each coefficient differs significantly from zero under the Wald test.
# fit a generalized linear model for each gene
gene_fits <- fit_models(cds_subset, model_formula_str = "~cell_type + batch")
# extract a table of coefficients
fit_coefs <- coefficient_table(gene_fits)
# preview the content of 'fit_coefs'
head(fit_coefs)
If you inspect the fit_coefs
object, you will notice that the table includes one row for each term of each gene’s model. We generally don’t care about the intercept term, so we can filter it out. In this way, we will be able to control for the chosen factors. To focus on only one variable, you have to check the term
column in the fit_coefs
and pass this as an argument for filtering. Then, you should also filter the results with q_value < 0.05 to control the false discovery rate.
# filter out Intercept term
no_intercept_coefs <- fit_coefs %>% dplyr::filter(term != "(Intercept)")
# extract results for DP-M1 cells only
DP_M1_coefs <- fit_coefs %>% dplyr::filter(term == "cell_typeDP-M1")
# control the false discovery rate and choose only several variables to store
DP_M1_coefs_filtered <- DP_M1_coefs %>% dplyr::filter (q_value < 0.05) %>%
dplyr::select(gene_short_name, term, q_value, estimate)
# view the resulting table
DP_M1_coefs_filtered
The resulting table shows the genes that differ depending on the chosen term. Maybe this function is not very helpful in the case of our dataset, but may be useful when analysing unannotated data or choosing another term from colData()
.
Graph-autocorrelation analysis - advanced
Alongside regression analysis, Monocle also provides another way of finding genes that vary between groups of cells. The function graph_test()
uses a statistic from spatial autocorrelation analysis called Moran’s I, which Cao et al. 2018 showed to be effective in finding genes that vary in single-cell RNA-seq datasets. Let’s try to perform this step on our full dataset (be patient!).
# run autocorrelation test
graph_test_res <- graph_test(cds_order, neighbor_graph="knn", cores=8)
The output data frame has a bunch of statistical values that you can use to rank the genes by, for example morans_I
column, which ranges from -1 to +1. A value of 0 indicates no effect, while +1 indicates perfect positive autocorrelation and suggests that nearby cells have very similar values of a gene’s expression.
We will now try to associate genes with clusters by grouping genes into modules that have similar patterns of expression. You can call find_gene_modules()
, which essentially runs UMAP on the genes (as opposed to the cells) and then groups them into modules using Louvain community analysis.
# get gene IDs for which q-value < 0.05
pr_deg_ids <- row.names(subset(graph_test_res, q_value < 0.05))
# group genes into modules
gene_module_df <- find_gene_modules(cds_order[pr_deg_ids,], resolution=1e-2)
Now we can show the aggregate expression of all genes in each module across all the clusters. Monocle provides a simple utility function called aggregate_gene_expression
for this purpose:
# aggregate expression of genes in each module
cell_group_df <- tibble::tibble(cell=row.names(colData(cds_order)),
cell_group=clusters(cds_order)[colnames(cds_order)])
agg_mat <- aggregate_gene_expression(cds_order, gene_module_df, cell_group_df)
row.names(agg_mat) <- stringr::str_c("Module ", row.names(agg_mat))
colnames(agg_mat) <- stringr::str_c("Cluster ", colnames(agg_mat))
We can now use this data to create a heatmap. Don’t worry if yours look a little bit different that the one shown in this tutorial. The general features should be maintained though.
# create a heatmap
pheatmap::pheatmap(agg_mat, cluster_rows=TRUE, cluster_cols=TRUE,
scale="column", clustering_method="ward.D2",
fontsize=6)
You can also visualise the modules using plot_cells()
function. We’ve chosen some modules to see how the differences on a heatmap correlate with the expression shown on our starting plot.
QuestionWhich modules to plot?
This is totally up to you! It might be the case that you got different numbering of modules, so then using the numbers specified in the code below won’t make much sense. Just look at your heatmap, compare the differences between modules and think which ones would be the most interesing to visualise.
# see the chosen modules across the whole sample
plot_cells(cds_order,
genes=gene_module_df %>% dplyr::filter(module %in% c(40, 39, 36, 17)),
group_cells_by="cluster",
color_cells_by="cluster",
show_trajectory_graph=FALSE)
With the visualisation methods above, you can now come back to the generated data frame gene_module_df
, filter genes that belong to the module of interest and check their functions to get some more evidence for the correct biological interpretation.
3D plotting
Let’s have some fun at the end! That was quite a long and insightful analysis – you definitely deserve to look at some nice, rotatable, cool plot now!
Essentially the workflow is the same as we followed in two dimensions. The crucial part is to specify the dimensionality of the reduced space with the max_components
parameter in reduce_dimension()
function. The default is 2, but if we want to see our data in 3D, we will change that value to 3. From there, you can just repeat the next steps in 3D… or just reward yourself for completing this tutorial by toggling this beautiful 3D plot!
# reduce dimension to 3D
cds_3d <- reduce_dimension(cds_order, preprocess_method = 'Aligned', max_components = 3)
# see the resulting 3D plot
plot_cells_3d(cds_3d, color_cells_by="cell_type")
Export your data, figures, and notebook
Don’t forget to save and export your data! First, we will get Jupyter to see those as files.
Export plots
If you want to export your plot, you have to make sure that you assigned it to an object. For example, if you want to save the plot of cells in pseudotime, simply assign the function you used to generate this plot to an object. Here we call this object plot_pseudotime
, like so:
plot_pseudotime <- plot_cells(cds_order,
color_cells_by = "pseudotime",
label_cell_groups=FALSE,
label_leaves=FALSE,
label_branch_points=FALSE)
Then, if you want to save the plot as PDF:
pdf("plot_pseudotime.pdf") # open the graphical device and specify the directory and the name of the output pdf file
plot_pseudotime # specify the object that your plot is assigned to
dev.off() # close the graphical device
The procedure is very similar if you want to export the file as PNG (or analogically JPEG – just replace png with jpeg):
png("plot_pseudotime.png", # open the graphical device and specify the directory and the name of the output png file
width=600, height=400) # optionally you can specify the width and height of the final plot
plot_pseudotime # specify the object that your plot is assigned to
dev.off() # close the graphical device
However, it often happens that the quality of the exported PNG and JPEG files is not perfect. For best results, we recommend exporting to SVG:
svg("plot_pseudotime.svg") # open the graphical device and specify the directory and the name of the output svg file
plot_pseudotime # specify the object that your plot is assigned to
dev.off() # close the graphical device
You can do the same with any plot that you want to save! You will find the saved figures in the left panel of your JupyterLab. You can right-click on them and download directly onto your computer. You can also push them into your Galaxy history. To do so, you have to change Kernel to Python3 (either click on Kernel
-> Change Kernel...
in the upper left corner of your JupyterLab or click on the displayed current kernel in the upper right corner and change it).
Check in the upper right corner that selected kernel is Python3, and run the following:
put("marker_file.txt")
put("plot_pseudotime.pdf")
put("plot_pseudotime.png")
In this way you can push all the files you’ve saved into your Galaxy history. You can also do the same with this notebook. The cell below will only work if you haven’t changed the name of the notebook. If you renamed it, simply type its new name in the parenthesis.
put("single-cell-scrna-case_monocle3-rstudio.ipynb")
Now you can go check your Galaxy history to make sure your files have all made it back into your Galaxy history.
After Jupyter
Congratulations! You’ve made it through Jupyter!
Hands-on: Closing JupyterLab
- Click User: Active Interactive Tools
- Tick the box of your Jupyter Interactive Tool, and click Stop
If you want to run this notebook again, or share it with others, it now exists in your history. You can use this ‘finished’ version just the same way as you downloaded the directions file and uploaded it into the Jupyter environment.
Conclusion
If you’re following the Case Study tutorials from the beginning, you have already experienced what it’s like to analyse and question a dataset, potentially without clear cut-offs or clear answers. The Monocle in Galaxy tutorial was focused more on explaining each step of the trajectory analysis and interpreting the results in the biological context. The current tutorial aims at showing the variety of methods that can be used when Galaxy’s Monocle tools are not enough. It shows the potential of batch correction, differential expression analysis and flexibility when using different functions. It’s also a guide for those who would like to understand what is happening ‘behind the scenes’ when clicking on Galaxy buttons.