sleuth icon indicating copy to clipboard operation
sleuth copied to clipboard

Allow for filter functions that operate on the whole matrix rather than by row

Open byoo opened this issue 6 years ago • 6 comments

Hello,

I have RNA sequencing data prepared using rRNA depletion and would like to conduct gene-level differential expression analysis using p-value aggregation. In doing so, I attempted to exclude non coding RNA such as rRNA by specifying target_ids of interest as follows: sleuth_prep(..., filter_target_id=protein_coding_target_ids)

However, it seems to lead to an error in calling sleuth_results

sleuth_results(so_g, 'reduced:full', 'lrt', show_all=T)
Error in sleuth_results(so_g, "reduced:full", "lrt", show_all = T) : 
  The provided weighting function for the mean observations results in negative values, which are not allowed for the lancaster method.

It runs successfully without the filtering. Would you advise how the issue can be resolved? I also wonder if the transcript filtering is correct.

Thank you!

byoo avatar Jun 15 '18 15:06 byoo

Ok, it looks that specifying filter_target_id made sleuth_prep not filter transcripts based on counts (basic_filter), which is necessary. Calling basic_filter resolved the error.

BTW, my implementation seems less optimal. I couldn't find a way to access transcript_id in filter_fun of sleuth_prep, so I had to run sleuth_prep twice to apply both filters as follows:

biotypefilter <- protein_coding_target_ids
so <- sleuth_prep(full_data, target_mapping=t2g, extra_bootstrap_summary=T,
                  num_cores=1)

txfilter <- so$filter_df$target_id[so$filter_df$target_id %in% biotypefilter]

so <- sleuth_prep(full_data, target_mapping=t2g, extra_bootstrap_summary=T,
                  num_cores=1, filter_target_id=txfilter,
                  pval_aggregate=T, aggregation_column='ens_gene')
...

Have I resolved it in a correct way? I would appreciate it if you would give any advice. Thanks!

byoo avatar Jun 15 '18 17:06 byoo

Hi @byoo,

The new filter_target_id option is working as described: you provide it a list of target_ids and it will only filter those exact IDs and nothing else. Since you are using a list of target_ids, the filter_fun is NULL. The error is occurring because you included low-abundance features by filtering all protein-coding genes, and any feature with an average (normalized) estimated count of less than 0.5 will yield a negative logarithm, and negative weights are not allowed for the p-value aggregation method.

Your way is correct. A slightly better way to do it is the following:

library(sleuth)
biotypefilter <- protein_coding_target_ids
so <- sleuth_prep(full_data, normalize = FALSE) ## this just reads in the raw data
counts <- sleuth_to_matrix(so, "obs_raw", "est_counts") ## this gets a matrix of counts
filter <- apply(counts, 1, basic_filter) ## this applies the basic filter to the counts
txfilter <- biotypefilter[which(biotypefilter %in% names(filter)[filter])] ## this gives you the filter

so <- sleuth_prep(full_data, target_mapping=t2g, extra_bootstrap_summary=T,
                  num_cores=1, filter_target_id=txfilter,
                  pval_aggregate=T, aggregation_column='ens_gene')

That being said, it may be worth it to reconfigure filtering to take the full matrix instead of forcing the filter function to operate by row. What do you think about that @pimentel?

warrenmcg avatar Jun 16 '18 00:06 warrenmcg

Thanks a lot for your kind help, @warrenmcg!

byoo avatar Jun 16 '18 03:06 byoo

Hi, I don't know if this is exactly the same issue but I'm getting the same message and I guess there is something that I'm missing.

I'm trying to use filter_fun with the new settings pval_aggregate= TRUE, aggregation_column='ens_gene' in sleuth 0.30.

I want to use filter_fun because I have a study with two genotypes and two treatments, so I have four groups. I didn't have any issue when using sleuth_0.29.

The problem that I get is when I do the LRT testing:

This is my code:

> my_filter= function (row, min_reads = 5, min_prop = 0.20)
{   
    mean(row >= min_reads) >= min_prop
}

> so= sleuth_prep( s2c, 
	target_mapping= t2g, 
	read_bootstrap_tpm= TRUE, 
	extra_bootstrap_summary= TRUE,
	pval_aggregate= TRUE,
	aggregation_column='ens_gene',
	filter_fun=my_filter)

> so= sleuth_fit( so, ~genotype+sex+tto, 'full')
fitting measurement error models
shrinkage estimation
3 NA values were found during variance shrinkage estimation due to mean observation values outside of the range used for the LOESS fit.
The LOESS fit will be repeated using exact computation of the fitted surface to extrapolate the missing values.
These are the target ids with NA values: ENSMUST00000080834.14, ENSMUST00000112010.8, ENSMUST00000127027.1
computing variance of betas
> so= sleuth_fit( so, ~genotype+sex, 'reduced')
fitting measurement error models
shrinkage estimation
3 NA values were found during variance shrinkage estimation due to mean observation values outside of the range used for the LOESS fit.
The LOESS fit will be repeated using exact computation of the fitted surface to extrapolate the missing values.
These are the target ids with NA values: ENSMUST00000080834.14, ENSMUST00000112010.8, ENSMUST00000127027.1
computing variance of betas
> so= sleuth_lrt( so, 'reduced', 'full')
> lrt_res=sleuth_results( so, 'reduced:full', 'lrt', show_all= FALSE)
Error in sleuth_results(so, "reduced:full", "lrt", show_all = FALSE) : 
  The provided weighting function for the mean observations results in negative values, which are not allowed for the lancaster method.

I hope you can help me with this! Thanks

m3lorra avatar Feb 11 '19 17:02 m3lorra

Is there any update on this? I have just finished running a large analysis learning just now at the last step that I need to start over with the filtering. I defined a transcript as expressed if it has six or more raw counts and excluded transcripts not expressed in at least 20 % of cases and 20% of controls. Additionaly, I filtered by biotype since my kallisto index contains rRNAs etc that I want to account for during pseudoalignment but do not care about regarding differential expression. Something like a warning during sleuth_prep concerning too low expressed genes would have been useful already.

mschilli87 avatar Dec 02 '21 18:12 mschilli87

@warrenmcg:

Can you confirm that the following is equivalent to your suggestion?

#[...]
so <- sleuth_prep(full_data, read_bootstraps = FALSE, normalize = FALSE) ## read/process as little as nessecary
filter <- so$filter_bool ## no need to get the count matrix and run the basic filter again
#[...]

Any trick to speed things up by filtering down the existing in RAM sleuth object instead of re-reading the data from disk?

Or even better: Why not issue a warning for neg. values during p-value aggregation and return NA results for the corresponding genes instead of failing the entire analysis? AFAICT, even with the basic filter applied, a mean normalized expression below 0.5 could still arise depending on the dataset. In such a case, the user is left with an error message referring to weights they never explicitly specified making this hard to debug. And even with the information provided here, finding an appropriate filter to remove the (unknown) offending targets could be more complecated than it needs to be.

mschilli87 avatar Dec 02 '21 20:12 mschilli87