Proteomics is the large-scale study of proteins. Proteins are vital parts of living organisms, with many functions. The term proteomics was coined in 1997, in analogy to genomics, the study of the genome. The word proteome is a portmanteau of protein and genome, and was coined by Marc Wilkins in 1994 while he was a Ph.D. student at Macquarie University. Macquarie University also founded the first dedicated proteomics laboratory in 1995.
The proteome
is the entire set of proteins that is produced or modified by an
organism or system. Proteomics has enabled the identification of ever
increasing numbers of protein. This varies with time and distinct
requirements, or stresses, that a cell or organism undergoes.
Proteomics is an interdisciplinary domain that has benefitted greatly
from the genetic information of various genome projects, including the Human Genome Project.
It covers the exploration of proteomes from the overall level of
protein composition, structure, and activity. It is an important
component of functional genomics.
Proteomics generally refers to the large-scale
experimental analysis of proteins and proteomes, but often is used
specifically to refer to protein purification and mass spectrometry.
Complexity of the problem
After genomics and transcriptomics,
proteomics is the next step in the study of biological systems. It is
more complicated than genomics because an organism's genome is more or
less constant, whereas proteomes differ from cell to cell and from time
to time. Distinct genes are expressed in different cell types, which
means that even the basic set of proteins that are produced in a cell
needs to be identified.
In the past this phenomenon was assessed by RNA analysis, but it was found to lack correlation with protein content. Now it is known that mRNA is not always translated into protein,
and the amount of protein produced for a given amount of mRNA depends
on the gene it is transcribed from and on the current physiological
state of the cell. Proteomics confirms the presence of the protein and
provides a direct measure of the quantity present.
Post-translational modifications
Not
only does the translation from mRNA cause differences, but many
proteins also are subjected to a wide variety of chemical modifications
after translation. Many of these post-translational modifications are
critical to the protein's function.
Phosphorylation
One such modification is phosphorylation, which happens to many enzymes and structural proteins in the process of cell signaling. The addition of a phosphate to particular amino acids—most commonly serine and threonine mediated by serine-threonine kinases, or more rarely tyrosine mediated by tyrosine kinases—causes
a protein to become a target for binding or interacting with a distinct
set of other proteins that recognize the phosphorylated domain.
Because protein phosphorylation is one of the most-studied
protein modifications, many "proteomic" efforts are geared to
determining the set of phosphorylated proteins in a particular cell or
tissue-type under particular circumstances. This alerts the scientist to
the signaling pathways that may be active in that instance.
Ubiquitination
Ubiquitin is a small protein that may be affixed to certain protein substrates by enzymes called E3 ubiquitin ligases.
Determining which proteins are poly-ubiquitinated helps understand how
protein pathways are regulated. This is, therefore, an additional
legitimate "proteomic" study. Similarly, once a researcher determines
which substrates are ubiquitinated by each ligase, determining the set
of ligases expressed in a particular cell type is helpful.
Additional modifications
In addition to phosphorylation and ubiquitination, proteins may be subjected to (among others) methylation, acetylation, glycosylation, oxidation, and nitrosylation.
Some proteins undergo all these modifications, often in time-dependent
combinations. This illustrates the potential complexity of studying
protein structure and function.
Distinct proteins are made under distinct settings
A cell may make different sets of proteins at different times or under different conditions, for example during development, cellular differentiation, cell cycle, or carcinogenesis.
Further increasing proteome complexity, as mentioned, most proteins are
able to undergo a wide range of post-translational modifications.
Therefore, a "proteomics" study may become complex very quickly,
even if the topic of study is restricted. In more ambitious settings,
such as when a biomarker
for a specific cancer subtype is sought, the proteomics scientist might
elect to study multiple blood serum samples from multiple cancer
patients to minimise confounding factors and account for experimental
noise. Thus, complicated experimental designs are sometimes necessary to account for the dynamic complexity of the proteome.
Limitations of genomics and proteomics studies
Proteomics gives a different level of understanding than genomics for many reasons:
- the level of transcription of a gene gives only a rough estimate of its level of translation into a protein. An mRNA produced in abundance may be degraded rapidly or translated inefficiently, resulting in a small amount of protein.
- as mentioned above many proteins experience post-translational modifications that profoundly affect their activities; for example, some proteins are not active until they become phosphorylated. Methods such as phosphoproteomics and glycoproteomics are used to study post-translational modifications.
- many transcripts give rise to more than one protein, through alternative splicing or alternative post-translational modifications.
- many proteins form complexes with other proteins or RNA molecules, and only function in the presence of these other molecules.
- protein degradation rate plays an important role in protein content.
Reproducibility. One major factor affecting reproducibility in
proteomics experiments is the simultaneous elution of many more
peptides than mass spectrometers can measure. This causes stochastic differences between experiments due to data-dependent acquisition
of tryptic peptides. Although early large-scale shotgun proteomics
analyses showed considerable variability between laboratories,
presumably due in part to technical and experimental differences
between laboratories, reproducibility has been improved in more recent
mass spectrometry analysis, particularly on the protein level and using
Orbitrap mass spectrometers. Notably, targeted proteomics
shows increased reproducibility and repeatability compared with shotgun
methods, although at the expense of data density and effectiveness.
Methods of studying proteins
In proteomics, there are multiple methods to study proteins. Generally, proteins may be detected by using either antibodies (immunoassays) or
mass spectrometry. If a complex biological sample is analyzed, either a very specific antibody needs to be used in quantitative dot blot analysis (qdb), or biochemical separation then needs to be used before the detection step, as there are
too many analytes in the sample to perform accurate detection and
quantification.
Protein detection with antibodies (immunoassays)
Antibodies to particular proteins, or to their modified forms, have been used in biochemistry and cell biology
studies. These are among the most common tools used by molecular
biologists today. There are several specific techniques and protocols
that use antibodies for protein detection. The enzyme-linked immunosorbent assay (ELISA) has been used for decades to detect and quantitatively measure proteins in samples. The Western blot
may be used for detection and quantification of individual proteins,
where in an initial step, a complex protein mixture is separated using SDS-PAGE and then the protein of interest is identified using an antibody.
Modified proteins may be studied by developing an antibody specific to that modification. For example, there are antibodies that only recognize certain proteins when they are tyrosine-phosphorylated,
they are known as phospho-specific antibodies. Also, there are
antibodies specific to other modifications. These may be used to
determine the set of proteins that have undergone the modification of
interest.
Disease detection at the molecular level is driving the emerging
revolution of early diagnosis and treatment. A challenge facing the
field is that protein biomarkers for early diagnosis may be present in
very low abundance. The lower limit of detection with conventional
immunoassay technology is the upper femtomolar range (10(-13) M).
Digital immunoassay technology has improved detection sensitivity three
logs, to the attomolar range (10(-16) M). This capability has the
potential to open new advances in diagnostics and therapeutics, but such
technologies have been relegated to manual procedures that are not well
suited for efficient routine use.
Antibody-free protein detection
While
protein detection with antibodies is still very common in molecular
biology, other methods have been developed also, that do not rely on an
antibody. These methods offer various advantages, for instance they
often are able to determine the sequence of a protein or peptide, they
may have higher throughput than antibody-based, and they sometimes can
identify and quantify proteins for which no antibody exists.
Detection methods
One of the earliest methods for protein analysis has been Edman degradation (introduced in 1967) where a single peptide
is subjected to multiple steps of chemical degradation to resolve its
sequence. These early methods have mostly been supplanted by
technologies that offer higher throughput.
More recently implemented methods use mass spectrometry-based
techniques, a development that was made possible by the discovery of
"soft ionization" methods developed in the 1980s, such as matrix-assisted laser desorption/ionization (MALDI) and electrospray ionization (ESI). These methods gave rise to the top-down and the bottom-up proteomics workflows where often additional separation is performed before analysis.
Separation methods
For the analysis of complex biological samples, a reduction of sample complexity is required. This may be performed off-line by one-dimensional or two-dimensional
separation. More recently, on-line methods have been developed where
individual peptides (in bottom-up proteomics approaches) are separated
using Reversed-phase chromatography and then, directly ionized using ESI; the direct coupling of separation and analysis explains the term "on-line" analysis.
Hybrid technologies
There
are several hybrid technologies that use antibody-based purification of
individual analytes and then perform mass spectrometric analysis for
identification and quantification. Examples of these methods are the
MSIA (mass spectrometric immunoassay), developed by Randall Nelson in 1995, and the SISCAPA (Stable Isotope Standard Capture with Anti-Peptide Antibodies) method, introduced by Leigh Anderson in 2004.
Current research methodologies
Fluorescence two-dimensional differential gel electrophoresis (2-D DIGE)
may be used to quantify variation in the 2-D DIGE process and establish
statistically valid thresholds for assigning quantitative changes
between samples.
Comparative proteomic analysis may reveal the role of proteins in
complex biological systems, including reproduction. For example,
treatment with the insecticide triazophos causes an increase in the
content of brown planthopper (Nilaparvata lugens (Stål)) male
accessory gland proteins (Acps) that may be transferred to females via
mating, causing an increase in fecundity (i.e. birth rate) of females.
To identify changes in the types of accessory gland proteins (Acps) and
reproductive proteins that mated female planthoppers received from male
planthoppers, researchers conducted a comparative proteomic analysis of
mated N. lugens females. The results indicated that these proteins participate in the reproductive process of N. lugens adult females and males.
Proteome analysis of Arabidopsis peroxisomes has been established as the major unbiased approach for identifying new peroxisomal proteins on a large scale.
There are many approaches to characterizing the human proteome,
which is estimated to contain between 20,000 and 25,000 non-redundant
proteins. The number of unique protein species likely will increase by
between 50,000 and 500,000 due to RNA splicing and proteolysis events,
and when post-translational modification also are considered, the total
number of unique human proteins is estimated to range in the low
millions.
In addition, the first promising attempts to decipher the proteome of animal tumors have recently been reported.
This method used as a functional method in Macrobrachium rosenbergii protein profiling.
High-throughput proteomic technologies
Proteomics
has steadily gained momentum over the past decade with the evolution of
several approaches. Few of these are new, and others build on
traditional methods. Mass spectrometry-based methods and micro arrays
are the most common technologies for large-scale study of proteins.
Mass spectrometry and protein profiling
There are two mass spectrometry-based methods currently used for
protein profiling. The more established and widespread method uses high
resolution, two-dimensional electrophoresis to separate proteins from
different samples in parallel, followed by selection and staining of
differentially expressed proteins to be identified by mass spectrometry.
Despite the advances in 2-DE and its maturity, it has its limits as
well.
The central concern is the inability to resolve all the proteins within a
sample, given their dramatic range in expression level and differing
properties.
The second quantitative approach uses stable isotope tags to
differentially label proteins from two different complex mixtures. Here,
the proteins within a complex mixture are labeled isotopically first,
and then digested to yield labeled peptides. The labeled mixtures are
then combined, the peptides separated by multidimensional liquid
chromatography and analyzed by tandem mass spectrometry. Isotope coded
affinity tag (ICAT) reagents are the widely used isotope tags. In this
method, the cysteine residues of proteins get covalently attached to the
ICAT reagent, thereby reducing the complexity of the mixtures omitting
the non-cysteine residues.
Quantitative proteomics using stable isotopic tagging is an
increasingly useful tool in modern development. Firstly, chemical
reactions have been used to introduce tags into specific sites or
proteins for the purpose of probing specific protein functionalities.
The isolation of phosphorylated peptides has been achieved using
isotopic labeling and selective chemistries to capture the fraction of
protein among the complex mixture. Secondly, the ICAT technology was
used to differentiate between partially purified or purified
macromolecular complexes such as large RNA polymerase II pre-initiation
complex and the proteins complexed with yeast transcription factor.
Thirdly, ICAT labeling was recently combined with chromatin isolation to
identify and quantify chromatin-associated proteins. Finally ICAT
reagents are useful for proteomic profiling of cellular organelles and
specific cellular fractions.
Another quantitative approach is the Accurate Mass and Time (AMT) tag approach developed by Richard D. Smith and coworkers at Pacific Northwest National Laboratory.
In this approach, increased throughput and sensitivity is achieved by
avoiding the need for tandem mass spectrometry, and making use of
precisely determined separation time information and highly accurate
mass determinations for peptide and protein identifications.
Protein chips
Balancing
the use of mass spectrometers in proteomics and in medicine is the use
of protein micro arrays. The aim behind protein micro arrays is to print
thousands of protein detecting features for the interrogation of
biological samples. Antibody arrays are an example in which a host of
different antibodies are arrayed to detect their respective antigens
from a sample of human blood. Another approach is the arraying of
multiple protein types for the study of properties like protein-DNA,
protein-protein and protein-ligand interactions. Ideally, the functional
proteomic arrays would contain the entire complement of the proteins of
a given organism. The first version of such arrays consisted of 5000
purified proteins from yeast deposited onto glass microscopic slides.
Despite the success of first chip, it was a greater challenge for
protein arrays to be implemented. Proteins are inherently much more
difficult to work with than DNA. They have a broad dynamic range, are
less stable than DNA and their structure is difficult to preserve on
glass slides, though they are essential for most assays. The global ICAT
technology has striking advantages over protein chip technologies.
Reverse-phased protein microarrays
This
is a promising and newer microarray application for the diagnosis,
study and treatment of complex diseases such as cancer. The technology
merges laser capture microdissection (LCM)
with micro array technology, to produce reverse phase protein
microarrays. In this type of microarrays, the whole collection of
protein themselves are immobilized with the intent of capturing various
stages of disease within an individual patient. When used with LCM,
reverse phase arrays can monitor the fluctuating state of proteome among
different cell population within a small area of human tissue. This is
useful for profiling the status of cellular signaling molecules, among a
cross section of tissue that includes both normal and cancerous cells.
This approach is useful in monitoring the status of key factors in
normal prostate epithelium and invasive prostate cancer tissues. LCM
then dissects these tissue and protein lysates were arrayed onto
nitrocellulose slides, which were probed with specific antibodies. This
method can track all kinds of molecular events and can compare diseased
and healthy tissues within the same patient enabling the development of
treatment strategies and diagnosis. The ability to acquire proteomics
snapshots of neighboring cell populations, using reverse phase
microarrays in conjunction with LCM has a number of applications beyond
the study of tumors. The approach can provide insights into normal
physiology and pathology of all the tissues and is invaluable for
characterizing developmental processes and anomalies.
Practical applications of proteomics
One
major development to come from the study of human genes and proteins
has been the identification of potential new drugs for the treatment of
disease. This relies on genome and proteome information to identify
proteins associated with a disease, which computer software can then use
as targets for new drugs. For example, if a certain protein is
implicated in a disease, its 3D structure provides the information to
design drugs to interfere with the action of the protein. A molecule
that fits the active site of an enzyme, but cannot be released by the
enzyme, inactivates the enzyme. This is the basis of new drug-discovery
tools, which aim to find new drugs to inactivate proteins involved in
disease. As genetic differences among individuals are found, researchers
expect to use these techniques to develop personalized drugs that are
more effective for the individual.
Proteomics is also used to reveal complex plant-insect
interactions that help identify candidate genes involved in the
defensive response of plants to herbivory.
Interaction proteomics and protein networks
Interaction
proteomics is the analysis of protein interactions from scales of
binary interactions to proteome- or network-wide. Most proteins
function via protein–protein interactions, and one goal of interaction proteomics is to identify binary protein interactions, protein complexes, and interactomes.
Several methods are available to probe protein–protein interactions. While the most traditional method is yeast two-hybrid analysis, a powerful emerging method is affinity purification followed by protein mass spectrometry using tagged protein baits. Other methods include surface plasmon resonance (SPR), protein microarrays, dual polarisation interferometry, microscale thermophoresis and experimental methods such as phage display and in silico computational methods.
Knowledge of protein-protein interactions is especially useful in regard to biological networks and systems biology, for example in cell signaling cascades and gene regulatory networks (GRNs, where knowledge of protein-DNA interactions
is also informative). Proteome-wide analysis of protein interactions,
and integration of these interaction patterns into larger biological networks, is crucial towards understanding systems-level biology.
Expression proteomics
Expression proteomics includes the analysis of protein expression
at larger scale. It helps identify main proteins in a particular
sample, and those proteins differentially expressed in related
samples—such as diseased vs. healthy tissue. If a protein is found only
in a diseased sample then it can be a useful drug target or diagnostic
marker. Proteins with same or similar expression profiles may also be
functionally related. There are technologies such as 2D-PAGE and mass
spectrometry that are used in expression proteomics.
Biomarkers
The National Institutes of Health
has defined a biomarker as “a characteristic that is objectively
measured and evaluated as an indicator of normal biological processes,
pathogenic processes, or pharmacologic responses to a therapeutic
intervention.”
Understanding the proteome, the structure and function of each
protein and the complexities of protein–protein interactions is critical
for developing the most effective diagnostic techniques and disease
treatments in the future. For example, proteomics is highly useful in
identification of candidate biomarkers (proteins in body fluids that are
of value for diagnosis), identification of the bacterial antigens that
are targeted by the immune response, and identification of possible
immunohistochemistry markers of infectious or neoplastic diseases.
An interesting use of proteomics is using specific protein
biomarkers to diagnose disease. A number of techniques allow to test for
proteins produced during a particular disease, which helps to diagnose
the disease quickly. Techniques include western blot, immunohistochemical staining, enzyme linked immunosorbent assay (ELISA) or mass spectrometry. Secretomics, a subfield of proteomics that studies secreted proteins
and secretion pathways using proteomic approaches, has recently emerged
as an important tool for the discovery of biomarkers of disease.
Proteogenomics
In proteogenomics, proteomic technologies such as mass spectrometry
are used for improving gene annotations. Parallel analysis of the
genome and the proteome facilitates discovery of post-translational
modifications and proteolytic events, especially when comparing multiple species (comparative proteogenomics).
Structural proteomics
Structural
proteomics includes the analysis of protein structures at large-scale.
It compares protein structures and helps identify functions of newly
discovered genes. The structural analysis also helps to understand that
where drugs bind to proteins and also show where proteins interact with
each other. This understanding is achieved using different technologies
such as X-ray crystallography and NMR spectroscopy.
Bioinformatics for proteomics (proteome informatics)
Much
proteomics data is collected with the help of high throughput
technologies such as mass spectrometry and microarray. It would often
take weeks or months to analyze the data and perform comparisons by
hand. For this reason, biologists and chemists are collaborating with
computer scientists and mathematicians to create programs and pipeline
to computationally analyze the protein data. Using bioinformatics
techniques, researchers are capable of faster analysis and data
storage. A good place to find lists of current programs and databases is
on the ExPASy
bioinformatics resource portal. The applications of
bioinformatics-based proteomics includes medicine, disease diagnosis,
biomarker identification, and many more.
Protein identification
Mass
spectrometry and microarray produce peptide fragmentation information
but do not give identification of specific proteins present in the
original sample. Due to the lack of specific protein identification,
past researchers were forced to decipher the peptide fragments
themselves. However, there are currently programs available for protein
identification. These programs take the peptide sequences output from
mass spectrometry and microarray and return information about matching
or similar proteins. This is done through algorithms implemented by the
program which perform alignments with proteins from known databases such
as UniProt and PROSITE to predict what proteins are in the sample with a degree of certainty.
Protein structure
The biomolecular structure
forms the 3D configuration of the protein. Understanding the protein's
structure aids in identification of the protein's interactions and
function. It used to be that the 3D structure of proteins could only be
determined using X-ray crystallography and NMR spectroscopy. As of 2017, Cryo-electron microscopy
is a leading technique, solving difficulties with crystallization (in
X-ray crystallography) and conformational ambiguity (in NMR); resolution
was 2.2Å as of 2015. Now, through bioinformatics, there are computer
programs that can in some cases predict and model the structure of
proteins. These programs use the chemical properties of amino acids and
structural properties of known proteins to predict the 3D model of
sample proteins. This also allows scientists to model protein
interactions on a larger scale. In addition, biomedical engineers are
developing methods to factor in the flexibility of protein structures to
make comparisons and predictions.
Post-translational modifications
Most programs available for protein analysis are not written for proteins that have undergone post-translational modifications.
Some programs will accept post-translational modifications to aid in
protein identification but then ignore the modification during further
protein analysis. It is important to account for these modifications
since they can affect the protein's structure. In turn, computational
analysis of post-translational modifications has gained the attention of
the scientific community. The current post-translational modification
programs are only predictive.
Chemists, biologists and computer scientists are working together to
create and introduce new pipelines that allow for analysis of
post-translational modifications that have been experimentally
identified for their effect on the protein's structure and function.
Computational methods in studying protein biomarkers
One
example of the use of bioinformatics and the use of computational
methods is the study of protein biomarkers. Computational predictive
models
have shown that extensive and diverse feto-maternal protein trafficking
occurs during pregnancy and can be readily detected non-invasively in
maternal whole blood. This computational approach circumvented a major
limitation, the abundance of maternal proteins interfering with the
detection of fetal proteins,
to fetal proteomic analysis of maternal blood. Computational models can
use fetal gene transcripts previously identified in maternal whole blood to create a comprehensive proteomic network of the term neonate.
Such work shows that the fetal proteins detected in pregnant woman’s
blood originate from a diverse group of tissues and organs from the
developing fetus. The proteomic networks contain many biomarkers
that are proxies for development and illustrate the potential clinical
application of this technology as a way to monitor normal and abnormal
fetal development.
An information theoretic framework has also been introduced for biomarker discovery, integrating biofluid and tissue information.
This new approach takes advantage of functional synergy between certain
biofluids and tissues with the potential for clinically significant
findings not possible if tissues and biofluids were considered
individually. By conceptualizing tissue-biofluid as information
channels, significant biofluid proxies can be identified and then used
for guided development of clinical diagnostics. Candidate biomarkers are
then predicted based on information transfer criteria across the
tissue-biofluid channels. Significant biofluid-tissue relationships can
be used to prioritize clinical validation of biomarkers.
Emerging trends in proteomics
A
number of emerging concepts have the potential to improve current
features of proteomics. Obtaining absolute quantification of proteins
and monitoring post-translational modifications are the two tasks that
impact the understanding of protein function in healthy and diseased
cells. For many cellular events, the protein concentrations do not
change; rather, their function is modulated by post-translational
modifications (PTM). Methods of monitoring PTM are an underdeveloped
area in proteomics. Selecting a particular subset of protein for
analysis substantially reduces protein complexity, making it
advantageous for diagnostic purposes where blood is the starting
material. Another important aspect of proteomics, yet not addressed, is
that proteomics methods should focus on studying proteins in the context
of the environment. The increasing use of chemical cross linkers,
introduced into living cells to fix protein-protein, protein-DNA and
other interactions, may ameliorate this problem partially. The challenge
is to identify suitable methods of preserving relevant interactions.
Another goal for studying protein is to develop more sophisticated
methods to image proteins and other molecules in living cells and real
time.
Proteomics for systems biology
Advances in quantitative proteomics would clearly enable more in-depth analysis of cellular systems. Biological systems are subject to a variety of perturbations (cell cycle, cellular differentiation, carcinogenesis, environment (biophysical), etc.). Transcriptional and translational
responses to these perturbations results in functional changes to the
proteome implicated in response to the stimulus. Therefore, describing
and quantifying proteome-wide changes in protein abundance is crucial
towards understanding biological phenomenon more holistically, on the level of the entire system. In this way, proteomics can be seen as complementary to genomics, transcriptomics, epigenomics, metabolomics, and other -omics approaches in integrative analyses attempting to define biological phenotypes more comprehensively. As an example, The Cancer Proteome Atlas
provides quantitative protein expression data for ~200 proteins in over
4,000 tumor samples with matched transcriptomic and genomic data from The Cancer Genome Atlas.
Similar datasets in other cell types, tissue types, and species,
particularly using deep shotgun mass spectrometry, will be an immensely
important resource for research in fields like cancer biology, developmental and stem cell biology, medicine, and evolutionary biology.
Human plasma proteome
Characterizing
the human plasma proteome has become a major goal in the proteomics
arena, but it is also the most challenging proteomes of all human
tissues.
It contains immunoglobulin, cytokines, protein hormones, and secreted
proteins indicative of infection on top of resident, hemostatic
proteins. It also contains tissue leakage proteins due to the blood
circulation through different tissues in the body. The blood thus
contains information on the physiological state of all tissues and,
combined with its accessibility, makes the blood proteome invaluable for
medical purposes. It is thought that characterizing the proteome of
blood plasma is a daunting challenge.
The depth of the plasma proteome encompassing a dynamic range of more than 1010
between the highest abundant protein (albumin) and the lowest (some
cytokines) and is thought to be one of the main challenges for
proteomics.
Temporal and spatial dynamics further complicate the study of human
plasma proteome. The turnover of some proteins is quite faster than
others and the protein content of an artery may substantially vary from
that of a vein. All these differences make even the simplest proteomic
task of cataloging the proteome seem out of reach. To tackle this
problem, priorities need to be established. Capturing the most
meaningful subset of proteins among the entire proteome to generate a
diagnostic tool is one such priority. Secondly, since cancer is
associated with enhanced glycosylation of proteins, methods that focus
on this part of proteins will also be useful. Again: multiparameter
analysis best reveals a pathological state. As these technologies
improve, the disease profiles should be continually related to
respective gene expression changes.
Due to the above-mentioned problems plasma proteomics remained
challenging. However, technological advancements and continuous
developments seem to result in a revival of plasma proteomics as it was
shown recently by a technology called plasma proteome profiling.
Due to such technologies researchers were able to investigate
inflammation processes in mice, the heritability of plasma proteomes as
well as to show the effect of such a common life style change like
weight loss on the plasma proteome.