Journal of Proteomics & Bioinformatics

Journal of Proteomics & Bioinformatics
Open Access

ISSN: 0974-276X

Research - (2024)Volume 17, Issue 2

Proteoscanner: A Full Proteomics Data Analysis and Visualization Open-Source Software

Ashraf Eltaher*
 
*Correspondence: Ashraf Eltaher, Department of Basic Research, Children's Cancer Hospital, Cairo, Egypt, Email: ,

Author info »

Abstract

Proteoscanner is an intuitive and versatile open-source software platform that offers a seamless and integrated approach for the analysis of proteomics data. With a focus on user-centric design, the software facilitates an end-to-end workflow that encompasses data preparation, pre-processing, statistical testing, visualization and biological interpretation. The incorporation of various statistical analyses and visualization tools within Proteoscanner allows for a sophisticated yet straightforward examination of complex datasets, ensuring that users can easily navigate through the analytical process. The software is an open-source, which allows full tweaking capabilities that can be summarized in transparency, flexibility, community-driven development, customizability, interoperability and long-term viability. It is compatible with different data array from current mass spectrometry equipment (Sciex and Thermo). This flexibility is further enhanced by the software's foundation in Python and R, allowing for cross-platform operability and exploiting the capabilities of the two languages. Proteoscanner has a standardized quality and efficiency in producing meaningful representations and generating comprehensive biological insights. Proteoscanner epitomizes the synthesis of user-friendly interface design with the analytical depth required for scientific inquiry. It stands as a testament to the software's commitment to facilitating scientific discovery, providing researchers with a tool that is not only comprehensive in its analytical capabilities but also intuitive in its use. This balance of sophistication and accessibility makes Proteoscanner an invaluable resource for the omics research community. The tool is available as an open-source downloadable executable file.

Keywords

Proteomics; Data analysis; Software tool; Statistics; Bioinformatics; Functional analysis

Introduction

Proteomics and metabolomics play a major rule in biological research and biomarker discovery and represent a key-part in understanding the integrated complex biological systems. However, the analysis of proteomics data presents significant obstacles due to the high dimensionality and complexity of the data. Nevertheless, mass spectrometry-based omics data analysis remains challenging due to the numerous widely available data analysis tools, the absence of unified standardized pipeline, the diversity of the generated data formats/sources as well as the data-dependent processing steps. These facts could lead to some confusion in selecting the most appropriate methods and mechanisms for data processing. A vast amount of software’s and web tools had been developed offering different analysis pipelines and data manipulation solutions, some of which provide a powerful robust pipeline but lack optionality while others offer a variety of analysis methodologies with powerful statistics but the quantity and/or quality of available visualization tools is insufficient. Proteoscanner is a user-friendly platform which allows the user to select from various options available for each pre/post-processing stage and/or compare different methods to reach the desired and most accurate analysis results. In addition, it provides precise statistical tests, vivid downloadable figures with publishable qualities and detailed customization options as well as biological interpretation capabilities, such as functional enrichment, over-representation analysis, gene ontologies and protein-protein interactions.

Mass Spectrometry (MS) is a powerful technique of analysis for both identification and quantification of a variety of analytes. When coupled with gas or liquid chromatographs, mass spectrometers expand analytical capabilities to serve more research applications including proteins which drives MS to be an essential analytical tool in the field of proteomics [1].

Modern proteomics experiments generate a vast array of data, presenting a significant challenge in terms of computation and analysis. In terms of complexity, proteomics data is almost>40 times more complex when compared to genomics data of the same test sample, this is due to the huge number of proteins available in comparison to the number of known genes. To address this challenge, sophisticated computational tools are indispensable for distilling these complex datasets into meaningful information. A plethora of attempts to introduce analysis tools that overcome the burdens were done and some great tool solutions were produced. Some of these tools are web-based which have the limitation of the permanent necessity of a powerful internet connection. Other tools are scripting tools which require a prior knowledge of coding and programming languages. Graphical User Interface (GUI) tools are usually user-friendly, but unfortunately, despite being very robust and high quality tools they are, in most cases, either a license-based paid tools or lack accessibility to source code which encounters some flexibility and interoperability issues.

Proteoscanner emerges as a critical solution, specifically designed to streamline the manipulation and analysis of LC/MS-based quantitative proteomics data. The software is an installable, majorly offline-based, free and open source GUI tool. It boasts a wide-scale platform that, although primarily intended for LC/MS proteomic data analysis, offers the versatility to extend to other omics types. This software stands out for its accessibility to users without prior knowledge in programming or statistics, making it a valuable tool for the broader scientific community. Offered as a downloadable package, Proteoscanner includes executable installation and application files along with comprehensive guides and dependencies to ensure ease of use.

In terms of its architecture, Proteoscanner is equipped to handle not only LC/MS-based quantitative proteomics data but also other omics datasets to a certain extent. It incorporates various R packages for analytical functions and utilizes the pyQt5 module for its user interface, enhancing the user experience with a thoughtfully designed dashboard [1].

The database integration of Proteoscanner is extensive, featuring connections with UniProt for protein information retrieval, gProfiler for gene ontology and pathways data from diverse sources and Search Tool for the Retrieval of Interacting Genes/ Proteins (STRING) for constructing networks of protein-protein interactions. Such integration underscores the software's capability in providing a comprehensive suite for biological data interpretation [2-4].

Data loading is designed to be intuitive, offering various options to accommodate different data formats and sources, from local data frames in common file formats to data outputs from established proteomics identification tools. Proteoscanner ensures that the quality of data is verified through a 'Quality check' function before proceeding, laying the groundwork for accurate statistical interpretation.

Proteoscanner's preprocessing panel offers an array of tools for data manipulation and preparation, addressing key steps such as normalization, filtration, imputation and transformation or scaling. These steps are critical for ensuring the integrity and comparability of data within proteomic, metabolomics and transcriptomic datasets. The software does not prescribe a single method but instead a bunch of different techniques, allowing analysts to select the most appropriate method based on the specifics of their data.

The data analysis component of Proteoscanner is built to handle the complexity of omics data, offering both descriptive and inferential statistical approaches. Recommendations for suitable statistical tests are provided based on the nature and distribution of the data and the biological context of the samples.

Visualization capabilities within Proteoscanner are robust and user-customizable, featuring a suite of 19 different graph plots that allow for both pre- and post-processing visualization, ensuring that users can gain comprehensive insights into their data with publishable quality outputs.

Lastly, functional enrichment analysis is another key feature, tapping into biological pathways’ databases to provide deep biological information on the functions and interactions of biological entities.

In summary, Proteoscanner is a multifaceted, user-friendly software designed to meet the diverse needs of the proteomics research community. It provides an integrated solution for data loading, preprocessing, statistical analysis and biological interpretation, all within a single platform. Its open architecture and comprehensive feature set make it an indispensable tool for researchers navigating the complexities of proteomics data.

Materials and Methods

Proteoscanner is a wide-scale software platform designed mainly-but not exclusively for Liquid Chromatography-Mass Spectrometry (LC/MS) proteomic data analysis where it can be extended to other omics types. No prior knowledge of programming or statistics is required-although preferable- as it is appropriate for non-scientific users as well. It is provided as a downloadable package containing an executable installation and application files in addition to installation guide and other dependency files. It is available as an open source tool for windows-based operating systems. Prerequisites for installation are windows 10 or higher as an operating system in which the latest version of both python (python 3.12 at the time of publication) and R (R-4.3.2) is installed. A documentation file for the software is also provided. Figure 1 illustrates the available dashboards and the workflow of Proteoscanner from data uploading towards data manipulation, analysis and visualization down to biological investigation (Figure 2).

Proteoscanner

Figure 1: Workflow sequence of Proteoscanner showing different data loading methods, then quality control, processing of data and the possible data analysis and visualization options.

workflow

Figure 2: Flowchart of Proteoscanner workflow with all possible options and decisions. Equation

The software is designed to deal mainly with LC/MS based quantitative proteomics data but it is valid partially to any other omics data. Several R packages are used in the analysis functions and the software dashboard was built using the python module pyQt5 graphical user interface designer. Example data integrated within the software to demonstrate its functionality are generated from real datasets obtained from in-house studies inside our laboratory and visualization is facilitated by the ggplot2 package in R, among other specialized packages [5-7].

Results

User interface

Proteoscanner provides a nice, simple and user-friendly interface which drives the user smoothly through the analysis workflow by the help of its six main panels (Figure 3).

interface

Figure 3: Proteoscanner user interface denoted with the six main panels.

Panel 1 for loading data.

Panel 2 for data preprocessing and preparation.

Panel 3 for data statistical analysis and visualizations.

Panel 4 for functional biological interpretation.

Panel 5 for displaying different datasets throughout the session.

Panel 6 for displaying generated figures.

Databases

Uniprot: Used for retrieval of protein information (name and length e.g., in case of Normalized Spectral Abundance Factor (NSAF) calculation in protein pilot software data).

gProfiler: Used for retrieval of gene ontology and pathways information from different sources (Gene ontology, Kyoto Encyclopedia of Genes and Genomes (KEGG), Reactome, Wiki pathways) [8-11].

STRING: Used to generate a network of protein-protein interactions.

Data loading

Proteoscanner offers a user-friendly interface for seamless data loading and management.

Loading of data to Proteoscanner dashboard is available in a variety of options:

Load a local collected data-frame of multiple samples in several formats (csv, txt, xls, xlsx, pkl or RData).

Load a ready available example data.

Load a user-designed data-frame (manual or automatic filling).

Load an output data from several quantitative proteomics identification tools such as Protein Pilot, Peptide Shaker, Proteome Discoverer, MaxQuant or any other tool [12-15].

The loaded data-frame must have a fixed layout (Samples in columns and variables in rows except the first row which should contain the group labels) (Table 1). The id column could be of any known database identifier type or even omics entity names.

Protein ID Sample 1 Sample 2 Sample 3 Sample 4 Sample 5 Sample 6 Sample 7 Sample 8
Label Group 1 Group 1 Group 1 Group 1 Group 2 Group 2 Group 2 Group 2
A0A5C2G0A4 4.49E-06 4.68E-06 3.59E-06 2.54E-06 8.36E-06 3.86E-06 6.29E-06 4.88E-06
A0A5C2GAC3 0.010632 0.004308 NA 0.00651 NA 0.01021 NA 0.028196
A0A5C2FXP0 0.00139376 0.00055973 0.00187288 0.00015628 0.00186689 0.00062168 NA NA
A0A5C2FVT6 0.01063176 0.00386875 0.000543 0.00639669 0.001988 0.01021037 0.0066 0.0281958
G3V2W1 NA 0.00013559 9.38E-05 0.00040619 0.00092277 0.00020645 NA NA
P01871 6.13E-05 0.00104496 9.45E-05 0.00247271 0.00052281 NA 0.0000832 6.13E-05
P02749 0.00048643 0.00083125 NA 0.00018177 0.00070993 NA NA 0.00048643

Table 1: Proteoscanner loaded data-frame correct layout.

The quality of the uploaded data must be checked first by ‘Quality check’ functionality before proceeding to further steps to ensure its validity for statistical interpretations. Qualified data is the one which is totally numeric (except for the first ID column) and fulfills the layout criteria mentioned above in addition to absence of non-numeric characters or a complete missing column or row.

Data preprocessing

The data processing panel in Proteoscanner provides a comprehensive set of tools to preprocess and edit proteomics data.

There are four main steps of data manipulation and preparation commonly utilized in the pipeline of proteomics along with metabolomics and transcriptomics datasets analysis.

Normalization

The primary aim of normalizing data is to unify the scale for a fair and proper comparison i.e., minimize the technical variations as possible while keeping biological variations.

Quantitative proteomics comparative studies depend primarily on accurate normalization, for which there are a massive variety of choices. Which is the best is based mainly on several criteria such as data type and distribution, Number of samples and the nature of observed values. Some Bioconductor R packages such as limma, Deseq2 and EdgeR are already available and widely used in genomic data normalization and differential analysis but indeed they are not recommended in other types of omics data due to the fact of high dispersibility of genomic data that usually follows negative binomial distribution which is actually not the case in other omics. However, Proteoscanner offers no single recommendation for a normalization method and the decision is left to the analyst to choose from eleven different techniques according to his personal preference which may be-sometimes-affected by the nature of the used data.

Total Sum Normalization (TSN): Normalize each peak in a sample through dividing by the sample’s total peaks [16].

Image

Median Normalization (MN): Normalize each peak in a sample through dividing by the median quotient which is the exponential of the natural logs of all samples’ medians after subtracting their mean from them [17].

Image

Probabilistic Quotient Normalization (PQN): Involves using the highest sample in terms of peaks as a reference sample. The normalization takes place by dividing each peak by the median of its division by the median of the reference sample. [18].

Image

Internal Standard Normalization (ISN): Each peak value is normalized against a selected reference feature where it is divided by the corresponding reference feature value [16].

Image

External Standard Normalization (ESN): Similar to ISN but differs in the reference feature being an external Quality Control (QC) feature of known concentrations in each sample [16].

Image

Compositional Data Approach Normalization (CODA): Data is Centered log-ratio (clr) transformed and its means are calculated then back-transformed to the compositional space using the inverse clr transformation where back-transformed means are used as a scaler denominator for normalization [19].

Image

Robust linear regression normalization (Rlr): This method assumes a linear relationship between the bias in the data and the magnitude of peaks. It fits a linear model by robust regression using an M estimator by Iterated Re-Weighted Least Squares (IWLS). Three variants of Rlr were explored called Rlr, RlrMA and RlrMA cyclic [20,21].

Local regression normalization (Loess): This methods-in contrast to Rlr- assumes a nonlinear relationship between bias and peaks magnitude in which data is first MA transformed then undergoes loess normalization either against the reference average sample (loessF) or pair-wise among all the samples (loessCyc) [21,22].

Variance Stabilization Normalization (VSN): Involves bringing the samples’ variance onto a same scale with a set of parametric transformations and maximum likelihood estimation [17,21].

Quantile Normalization (QN): Involves replacing each sample peak with the mean of the corresponding quantile to force the sample distribution to similarity [21,23].

Eigen MS normalization: This algorithm uses singular value decomposition to capture and remove biases from LC-MS peak intensity measurements. It has the advantage of handling the widespread missing measurements that are typical in LC-MS and has an overfitting-preventing approach. It is suitable to proteomics data [24].

Filtration

Filtering data is a very important aspect which greatly impacts the statistical analysis results. The inclusion of noise, inaccurate or unconfident features should be avoided in order to increase accuracy and minimize statistical type I error. Three different options available for filtering out datasets either per row/feature or per column/sample; a-filtration based on missing value percentage where the features/samples with a percentage of missing values exceeding the user-defined threshold are removed.

b-filtration based on Relative Standard Deviation (RSD) where the features/samples with higher RSD percentage than the userdefined threshold are removed.

c-filtration based on Interquartile Range (IQR) where values which lie outside the IQR of the feature/sample are considered as outliers and removed.

Imputation

In order to go further for differential analysis or statistics, missing values inside the dataset-if any- must be imputed i.e., replaced by real values which should be chosen precisely not to affect the overall data distribution and statistics. User can choose from a variety of commonly-used imputation techniques in omics according to the nature of his data and experimental conditions.

Lowest of Detection methods (LOD): Imputation takes place by using the lowest detected peak value across all the sample or a ratio of it [25].

Image

Left-censored Normal Distribution (lcND): In this approach, missing values are imputed using a random very low value from the data distribution with a mean equal to data mean minus 2.2 standard deviations and a standard deviation of almost third that of the original data.

Image

where X is a random variable with mean=xμ-2.2xσ and standard deviation=0.3xσ, where xμ, xσ are original data mean and standard deviation respectively [25].

Hot-Deck Imputation (HD): It is a method in which each missing value is replaced with available values from the same matrix (the donor) [26,27].

k-Nearest Neighbors imputation (kNN): This method imputes missing values in a dataset based on the values of its nearest neighbors [25,27].

Local Least Squares imputation (LLS): It is a local least squares regression model where first, k variables are selected by pearson, spearman or kendall correlation coefficients, then missing values are imputed by a linear combination of the k selected variables [25,28,29].

Random Forest imputation (RF): Better used in the case of mixed-type data. It uses a random forest trained on the observed values of a data matrix to predict the missing values [25,30].

Singular Value Decomposition imputation (SVD): It depends mainly on performing Principal Component Analysis (PCA) for the dataset. Imputation is based on the low-rank approximation obtained from the SVD of the dataset [25,29].

Bayesian Principal Component Analysis imputation (BPCA): Same as SVD but with different mechanism where it incorporates a statistical Bayesian approach allowing for uncertainty estimation [25,29].

Median imputation approaches: Including imputation with the feature median or a random value around with a range of 1% above or below (Median +/- 1%).

Transformation or scaling

Changing the scale or the range of the dataset usually aids in better visualization of data and better inspection of its real distribution. Transformation is-however- somehow more invasive to the dataset than scaling because it results in some alterations in data distribution, intra-distances and betweeness. Therefore, caution should be regarded when transforming data in order not to disturb the real variations or introduce some bias. Selecting the proper scaling method is mostly important for the choice of the suitable statistical test as well as helping for more informative plots. Proteoscanner offers users variable data scaling/transformation mechanisms where they can choose interchangeably among them or either compare with each other. Should the dataset be scaled or not, depends on several aspects including: data nature and distribution as well as the aim of the study, the statistical test to be held and also the personal preference of the statistician.

Image

Image

Image

Image

Image

Image

Image

These four technique are ordered if used simultaneously. If the user wants to change their order, he can perform a multi-step process in the desired order [31,32].

Data annotation

If the id column contains uniprot IDs, Protein names can be retrieved online at any step from uniprot database using uniprotR library in R through the ‘Retrieve protein names’ functionality in the menu bar. This helps for more informative results and visualizations [33].

Data normality

Before proceeding to statistical data analysis, it is strongly recommended to perform a data normality checking to reveal whether the data distribution follows preset assumptions or a known distribution (assumptions-based) or it is assumptions-free i.e., does not follow a known distribution. This type of test is called data parametricity check, where the data which follows one of the normal distributions is called ‘Parametric’ while if not, is called ‘Non-parametric’. This is a very important step in deriving the choice of the statistical tests which will be applied on the data. There are three well-known statistical tests used for checking the normality of the data; Shapiro Wilk’s, Kolmogorov Smirnov and Anderson Darling tests. Each of which provides a p-value threshold upon which the theory of data normal distribution could be accepted or rejected.

Data analysis

Statistics are the most crucial part of omics data analysis and the selection of the most convenient statistical approach is the cornerstone for biological data investigation. No matter the nature and characteristics of your proteomics dataset, you will probably find a suitable statistical test that correctly fits your experimental goal. Both types of statistics; either descriptive in which complete statistics of the data are provided or inferential including correlation, differential expression and fold change analyses are offered. Each of which is supplied with choice recommendations for the user-whenever possible- depending on data distribution and nature of biological samples.

Differential expression: Finding out the main players distinguishing between two or more different sets of observations or different groups of phenotypes is the ultimate goal of differential analysis which leads to uncovering the underlying biological backgrounds. This set of main players is described by being the most statistically significant entities between the tested groups. A variety of statistical tests commonly used in proteomics and/or metabolomics are provided along with recommendations for the suitability of usage for each and the most suitable one in the current case (Figures 4 and 5).

screenshot

Figure 4: A screenshot of available statistical tests for each data type with recommendation.

generated

Figure 5: A sort of plot types generated with Proteoscanner.

A significance threshold (p-value or adjusted p-value) is selected also by the user beyond which the statistical difference is significant.

Note: when having more than two groups to be tested, an Analysis of Variance (ANOVA) test is done-or one of its derivatives-followed by a post-hoc test to determine the pairwise differences among the groups (Table 2).

Accessions Chi-squared P-value False discovery rate Status Significant-pair Post-hoc p-value
A0A024R326 6.49580439 0.08982812 0.44716633 Non-significant - -
A0A024R4E5 7.5821903 0.05548382 0.34430369 Non-significant OS-C 0.035754919
A0A024R4M0 6.30571542 0.0976477 0.46400965 Non-significant - -
A0A087WTP3 0.04525253 0.99747422 0.99990532 Non-significant - -
A0A087WTZ5 0.92769437 0.81873989 0.99990532 Non-significant - -
A0A087WU08 0.23473868 0.97179553 0.99990532 Non-significant - -
A0A087X0X3 16.7268412 0.0008043 0.07286934 Significant OS-C , OS-F 0.0136620733146763, 0.00192430727976067
A0A087X2D0 5.76781589 0.12346942 0.51313436 Non-significant - -
A0A0A0MQT7 10.8793276 0.01239662 0.23215881 Significant F-C 0.005834613

Table 2: Non-parametric Analysis of Variance (ANOVA) with Post-hoc test result.

Correlation analysis: This analysis usually takes place to determine how strong are certain features or samples correlated together so that there is a possible potential biological relationship between them. There are two types of correlation tests (Pearson and Spearman’s correlation). Depending on the data parametricity, we could select either of them. The test results in a correlation coefficient and a p-value for each feature or sample pair. Correlation coefficient ranges from -1 (perfect inverse correlation) to 1 (perfect positive correlation) (Tables 3 and 4).

Sample 1 Sample 2 Correlation coefficient p-value
18C 3T 0.635396706 9.15E-10
3T 4T 0.704569225 4.44E-16
10C 5C -0.992156742 1.20E-06
18C 5C 0.198416206 0.003189453
18T 5C -0.992156742 1.20E-06
19T 5C 0.12267245 0.193509927
3T 5C -0.992156742 1.20E-06

Table 3: Pair-wise correlation data-frame between samples with p-value.

Accessions 10C 18C 5C 8C 18T 19T 3T 4T FC Significance
A0A024QZ30 8.10E-05 1.48E-05 0.00013 0.00013 0.000209 0.0001309 0.00097674 0.000130506 4.04706915 Significant
A0A024QZZ7 0.010631 0.0043 0.01011 0.028195 0.010117 0.006505 0.01011 0.01021 0.693939 Non-significant
A0A024R321 2.67E-05 2.25E-05 5.39E-05 5.39E-05 0.000102 5.39E-05 5.39E-05 8.14E-05 1.857964 Non-significant
A0A024R324 0.001393 0.00055 0.001013 0.001013 0.001872 0.00015 0.001866 0.000621 1.134931 Non-significant
A0A024R3Q0 0.001545 0.000827 0.00097 0.00097 0.00097 0.0010501 0.0009703 0.000883 0.898133 Non-significant

Table 4: Pair-wise correlation data-frame between samples with p-value.

Fold change analysis: This analysis is valid between two groups only where it calculates the ratio between the two groups’ means in each feature. If more than two groups, user can choose two of them to test. He can choose also how many folds the ratio should be to consider the feature as significantly expressed (Table 4).

Visualization

Effective visualization is crucial for interpreting proteomics data. A very powerful panel of versatile visualization plots is included which provide an outright and comprehensive insights to the data under investigation. The visualization can be held both pre or post-processing. There are 19 different graph plots produced by the help of R packages ggplot2 and others [8,34-46]. You can obtain various flexible, customizable and publishable high-quality graphs which provide an insightful biological image (Figure 5).

Univariate analysis: A collection univariate analysis graphs such as; bar, scatter, line and area plots is included with customization settings to fit different personal preferences. Other figures include heat maps, correlation plots, volcano plots as well as distribution plots (density and box plots).

Multivariate analysis: Proteoscanner contains a panel for the most commonly used multivariate analysis graphs such as PCA, Partial Least Squares-Discriminant Analysis (PLSDA), t-distributed Stochastic Neighbor Embedding (t-SNE) and dendrograms, each of which also contains very detailed customization settings. They are provided with tables of scores and loadings for each feature (table) along with different graphical representations of the analysis results (Figure 6).

Component

Figure 6: (A): Principal Component Analysis (PCA) plot; (B): Variable Importance Plots (VIP) for the highest 15 VIP score features; (C): Hierarchical clustering of samples within two groups based on Euclidean distance; (D): t-distributed Stochastic Neighbor Embedding (t-SNE) representation of 2 groups of samples. Equation

Networking plots: Networking plots such as; Sankey, chord, pie, donut and network plots show the networking structure and adjacency between entities of choice where the nodes represent the entities and the edges are the connections between them (Figures 7 and 8).

samples

Figure 7: (A): Donut plot showing the features detected in each of three samples with their ratios; (B): Network of the occurrence of features within each sample within two-groups samples. Equation

graphical

Figure 8: Different graphical representations possibilities for functional enrichment analysis and protein-protein interactions.

Functional enrichment

The biological interpretation is the utmost objective for omics data analysis. Accordingly, biological insights must be inferenced accurately for a biological experiment to be scientifically valid. Functional enrichment analysis provides a profound biological information of the whereabouts and functions of the biological entities. Databases such as KEGG, Reactome and Wiki pathways along with others contain information about biological pathways of different organisms [9-11].

Gene Ontology (GO) database provides a framework and set of concepts for describing the functions of gene products where it supports the computational representation of biological systems. It consists of three major components; biological process, molecular function and cellular component [8].

Over Representation Analysis (ORA) is a statistical method for objectively determining whether a gene set or pathway is more prevalent or relevantly enriched in a set of biological entities such as genes, proteins or metabolites than we expect by chance. This provides an image of the gene ontology components or pathways in which our biological variables of interest are significantly represented.

The software utilizes this method through its biological enrichment functionalities using gprofiler2 package in R (ref) which retain information from the GO, KEGG, Reactome and Wiki pathways databases. The results are displayed both in a table form (table) and also with various visualization possibilities (Figure 8) [7,9-11].

Protein-protein interaction is another functionality for retrieving and displaying information about interactions between proteins of interest both tabular and graphically as a network retrieved from STRING database for protein interactions through STRINGdb library in R [4].

Discussion

Proteoscanner was developed as a standalone open-source software using python and R programming languages. It provides a framework for proteomics -and other omics- data full analysis pipeline with a multi-optional, flexible and user friendly dashboard. Through case studies spanning a diverse set of biological samples, Proteoscanner has demonstrated its robustness and efficacy in generating clear biological insights. These case studies serve to underscore the software’s versatility in handling varying analytical demands and its capacity to deliver meaningful interpretations of biological data. While it offers recommendations for data processing, statistical test selection and visualization based on our team's statistical expertise, it intentionally leaves room for the researcher's discretion due to the absence of a single standardized pipeline for LC/MS proteomics. This flexibility empowers researchers to tailor the analysis to their unique requirements, selecting from a myriad of analytical pathways without constraint.

The dashboard consists of four main tool panels for data loading, data processing/statistical analysis, wealthy visualization plots possibilities and biological functional interpretation. This is in addition to two additional panels for the display of tables and figures. Each panel has a variety of buttons and functionalities to fit for different data formats, analysis approaches and methodologies. Researchers are allowed to apply a single function or a group of functions simultaneously in any order, select from a variety of choices according to their analyzed data type whilst a recommendation to the best choice is also offered. The software is freely available and is an open source platform which is suitable for both researchers and non-scientific individuals.

Conclusion

Proteoscanner is a comprehensive software package that empowers researchers to analyze and interpret proteomics data efficiently and accurately without the need for complicated calculations or exhausting coding. By integrating data loading, data processing, visualization plots and functional enrichment analysis, Proteoscanner provides a user-friendly interface for both novice and expert users. The software offers a range of advanced features and algorithms, enabling researchers to uncover valuable biological insights from complex proteomics experiments. Proteoscanner represents a significant advancement in the field of proteomics data analysis and holds great promise for accelerating discoveries in the field of proteomics research. Moreover, the software is available as an open-source which accounts for more adjustability and continuous improvement.

Acknowledgement

We would like to acknowledge the contributions of the proteomics and metabolomics research unit at children cancer hospital 57357 and everyone who contributed to open-source libraries used in Proteoscanner. We must also enlighten the effort of Renad Alaa and Ala’a Ahmed for participating with their effort and time in the software interface layout refinement and the software materials’ testing. This work would not have been possible without their valuable efforts.

References

Author Info

Ashraf Eltaher*
 
Department of Basic Research, Children's Cancer Hospital, Cairo, Egypt
 

Citation: Eltaher A (2024) Proteoscanner: A Full Proteomics Data Analysis and Visualization Open-Source Software. J Proteomics Bioinform. 17:665.

Received: 14-May-2024, Manuscript No. JPB-24-31343; Editor assigned: 16-May-2024, Pre QC No. JPB-24-31343 (PQ); Reviewed: 30-May-2024, QC No. JPB-24-31343; Revised: 06-Jun-2024, Manuscript No. JPB-24-31343 (R); Published: 13-Jun-2024 , DOI: 10.35248/0974-276X.24.17.665

Copyright: © 2024 Eltaher A. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited.

Top