Beruflich Dokumente
Kultur Dokumente
html
ABSTRACT
Bioinformatics is the new niche in biotechnology , the new investment opportunity for
venture capitalists, the new business opportunity for entrepreneurs; bioinformatics is
simply the new, new thing!
In an attempt to provide a simplistic definition of bioinformatics, scientists have
conjured up broad band of interpretations where software companies plan to declining
businesses by switching from writing algorithms for banking to diversify into
bioinformatics as a logical extension of their businesses. Bioinformatics is the symbiotic
relationship between computational end biological sciences.
Computational simulation of experimental biology is another important
application of bioinformatics which is aptly referred to as “in silico” testing. This is
perhaps an area that will expand in prolific way, given the need to obtain a grater degree
of predictability in animal and human clinical trails.
Quantum computing is the field where we use quantum physics and use that
system developed for solving general problems (even complex problems) which can be
solved by digital computer.
Almost all of today’s computers are based on simple turing theory and employ
boolean logic based on binary mathematics. Even ‘parallel” computers are really complex
turing engines employing multiple computing modules which deal with pieces of
incoming data(chunks bytes, instructions, etc.). There has been some research into
biological computing using enzymes or large-molecule systems as memory, shift
registers etc., but this has not proven to be very practical. Quantum computing is based
on a different physics than digital computing. Instead of having two(or three) states per
element like digital computers which are off on, or neither, quantum computers have all
the three states at the same time.
1 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
BIOINFORMATICS :
INTRODUCTION
The science of bioinformatics or computational biology is increasingly being used to improve the
quality of life as we know it.
Bioinformatics has developed out of the need to understand the code of life, DNA. Massive
DNA sequencing projects have evolved and added in the growth of the science of bioinformatics.
DNA the basic molecule of life directly controls the fundamental biology of life. It codes for
genes which code for proteins which determine the biological makeup of humans or any living
organism. It is variations and errors in the genomic DNA which ultimately define the likelihood
of developing diseases or resistance to these same disorders.
The ultimate goal of bioinformatics is to uncover the wealth of biological information hidden in
the mass of sequence data and obtain a clearer insight into the fundamental biology of organisms
and to use this information to enhance the standard of life for mankind.
It is being used now and in the foreseable future in the areas of molecular medicine to help
produce better and more customised medicines to prevent or cure diseases, it has environmental
benefits in, identifying waste cleanup bacteria and in agriculture it can be used for producing
high yield low maintenance crops. These are just a few of the many benefits bioinformatics will
help develop.
The genomic era has seen a massive explosion in the amount of biological information
available due to huge advances in the fields of molecular biology and genomics.
2 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
Biological applications:
Once all of the biological data is stored consistently and is easily available to the
scientific community, the requirement is then to provide methods for extracting the
meaningful information from the mass of data. Bioinformatic tools are software programs
that are designed to carry out this analysis step.
Factors that must be taken into consideration when designing these tools are:
• The end user (the biologist) may not be a frequent user of computer technology
• These software tools must be made available over the internet given the global
distribution of the scientific research community
The EBI provides a wide range of biological data analysis tools that fall into the
following four major categories:
• Homology and Similarity Tools
• Protein Function Analysis
• Structural Analysis
• Sequence Analysis
Homology and Similarity Tools :
Homologous sequences are sequences that are related by divergence from a common
ancestor. Thus the degree of similarity between two sequences can be measured while
their homology is a case of being either true of false.
This set of tools can be used to identify similarities between novel query sequences of
unknown structure and function and database sequences whose structure and function
have been elucidated.
Structural Analysis :
This set of tools allow you to compare structures with the known structure databases. The
function of a protein is more directly a consequence of its structure rather than its
sequence with structural homologs tending to share functions. The determination of a
protein's 2D/3D structure is crucial in the study of its function.
Sequence Analysis:
This set of tools allows you to carry out further, more detailed analysis on your query
sequence including evolutionary analysis, identification of mutations, hydropathy
regions, CpG islands and compositional biases. The identification of these and other
biological properties are all clues that aid the search to elucidate the specific function of
your sequence.
3 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
Biological databases:
Biological databases are archives of consistent data that are stored in a uniform and efficient
manner. These databases contain data from a broad spectrum of molecular biology areas.
Primary or archived databases contain information and annotation of DNA and protein
sequences, DNA and protein structures and DNA and protein expression profiles.
Secondary or derived databases are so called because they contain the results of analysis on the
primary resources including information on sequence patterns or motifs, variants and mutations
and evolutionary relationships. Information from the literature is contained in bibliographic
databases, such as Medline.
It is essential that these databases are easily accessible and that an intuitive query system is
provided to allow researchers to obtain very specific information on a particular biological
subject. The data should be provided in a clear, consistent manner with some visualisation tools
to aid biological interpretation.
Specialist databases for particular subjects have been set-up for example EMBL database for
nucleotide sequence data, Swiss-Prot protein database and PDB a 3D protein structure database.
Scientists also need to be able to integrate the information obtained from the underlying
heterogeneous databases in a sensible manner in order to be able to get a clear overview of their
biological subject. SRS (Sequence Retrieval System) is a powerful, querying tool provided by
the EBI that links information from more than 150 heterogeneous resources.
Sequence generation, and its subsequent storage, interpretation and analysis are entirely
computer dependent tasks. However, the molecular biology of an organism is a very
complex issue with research being carried out at different levels including the genome,
proteome, transcriptome and metabalome levels. Following on from the explosion in
volume of genomic data, similar increase in data have been observed in the fields of
proteomics, transcriptomics and metabalomics.
The first challenge facing the bioinformatics community today is the intelligent and
efficient storage of this mass of data. It is then their responsibility to provide easy and
4 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
reliable access to this data. The data itself is meaningless before analysis and the sheer
volume present makes it impossible for even a trained biologist to begin to interpret it
manually. Therefore, incisive computer tools must be developed to allow the extraction
of meaningful biological information.
There are three central biological processes around which bioinformatics tools must be
developed:
• DNA sequence determines protein sequence
• Protein sequence determines protein structure
• Protein structure determines protein function
The integration of information learned about these key biological processes should allow
us to achieve the long term goal of the complete understanding of the biology of
organisms.
APPLICATIONS
Real world applications of bioinformatics
The science of bioinformatics has many beneficial uses in the modern day world.
In the not too distant future, the potential for using genes themselves to treat disease may
become a reality. Gene therapy is the approach used to treat, cure or even prevent disease
by changing the expression of a persons genes. Currently, this field is in its infantile stage
with clinical trials for many different types of cancer and other diseases ongoing.
Increasing levels of carbon dioxide emission, mainly through the expanding use of fossil
fuels for energy, are thought to contribute to global climate change. Recently, the DOE
(Department of Energy, USA) launched a program to decrease atmospheric carbon
dioxide levels. One method of doing so is to study the genomes of microbes that use
carbon dioxide as their sole carbon source.
2.4 Biotechnology
The archaeon Archaeoglobus fulgidus and the bacterium Thermotoga maritima have
potential for practical applications in industry and government-funded environmental
remediation. These microorganisms thrive in water temperatures above the boiling point
and therefore may provide the DOE, the Department of Defence, and private companies
with heat-stable enzymes suitable for use in industrial processes,
Other industrially useful microbes include, Corynebacterium glutamicum which is of
high industrial interest as a research object because it is used by the chemical industry for
the biotechnological production of the amino acid lysine. The substance is employed as a
source of protein in animal nutrition. Lysine is one of the essential amino acids in animal
5 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
Lactococcus lactis is one of the most important micro-organisms involved in the dairy
industry, it is a non-pathogenic rod-shaped bacterium that is critical for manufacturing
dairy products like buttermilk, yogurt and cheese. This bacterium, Lactococcus lactis
ssp., is also used to prepare pickled vegetables, beer, wine, some breads and sausages and
other fermented foods. Researchers anticipate that understanding the physiology and
genetic make-up of this bacterium will prove invaluable for food manufacturers as well
as the pharmaceutical industry, which is exploring the capacity of L. lactis to serve as a
vehicle for delivering drugs.
The sequencing of genomes from all three domains of life, eukaryota, bacteria and
archaea means that evolutionary studies can be performed in a quest to determine the tree
of life and the last universal common ancestor.
For more interesting stories, check the archive at the Genome News Network .
For information on structural, functional and comparative analysis of genomes and genes
from a wide variety of organisms see The Institute of Genomic Research.
3. Agriculture
The sequencing of the genomes of plants and animals should have enormous benefits for
the agricultural community. Bioinformatic tools can be used to search for the genes
within these genomes and to elucidate their functions. This specific genetic knowledge
could then be used to produce stronger, more drought, disease and insect resistant crops
and improve the quality of livestock making them healthier, more disease resistant and
more productive.
6 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
5. Comparative studies
Analysing and comparing the genetic material of different species is an important method
for studying the functions of genes, the mechanisms of inherited diseases and species
evolution. Bioinformatics tools can be used to make comparisons between the numbers,
locations and biochemical functions of genes in different organisms.
Organisms that are suitable for use in experimental research are termed model organisms.
They have a number of properties that make them ideal for research purposes including
short life spans, rapid reproduction, being easy to handle, inexpensive and they can be
manipulated at the genetic level.
An example of a human model organism is the mouse. Mouse and human are very
closely related (>98%) and for the most part we see a one to one correspondence between
genes in the two species. Manipulation of the mouse at the molecular level and genome
comparisons between the two species can and is revealing detailed information on the
functions of human genes, the evolutionary relationship between the two species and the
molecular mechanisms of many human diseases.
7 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
Introduction :
The science of physics seeks to ask, and find precise answers to, basic questions about
why nature is as it is. Historically, the fundamental principles of physics have been
concerned with questions such as ``what are things made of?'' and ``why do things move
as they do?'' In his Principia, Newton gave very wide-ranging answers to some of these
questions. By showing that the same mathamatical equations could describe the motions
of everyday objects and of planets, he showed that an everyday object such as a tea pot is
made of essentially the same sort of stuff as a planet: the motions of both can be
described in terms of their mass and the forces acting on them. Nowadays we would say
that both move in such a way as to conserve energy and momentum. In this way, physics
allows us to abstract from nature concepts such as energy or momentum which always
obey fixed equations, although the same energy might be expressed in many different
ways: for example, an electron in the large electron-positron collider at CERN, Geneva,
can have the same kinetic energy as a slug on a lettuce leaf.
Another thing which can be expressed in many different ways is information. For
example, the two statements ``the quantum computer is very interesting'' and
``l'ordinateur quantique est tres interessant'' have something in common, although they
share no words. The thing they have in common is their information content. Essentially
the same information could be expressed in many other ways, for example by substituting
numbers for letters in a scheme such as a -> 97, b -> 98, c -> 99 and so on, in which case
the english version of the above statement becomes 116 104 101 32 113 117 97 110 116
117 109... . It is very significant that information can be expressed in different ways
without losing its essential nature, since this leads to the possibility of the automatic
manipulation of information: a machine need only be able to manipulate quite simple
things like integers in order to do surprisingly powerful information processing, from
document preparation to differential calculus, even to translating between human
languages. We are familiar with this now, because of the ubiquitous computer, but even
fifty years ago such a widespread significance of automated information processing was
not forseen.
However, there is one thing that all ways of expressing information must have in
common: they all use real physical things to do the job. Spoken words are conveyed by
air pressure fluctuations, written ones by arrangements of ink molecules on paper, even
thoughts depend on neurons (Landauer 1991). The rallying cry of the information
physicist is ``no information without physical representation!'' Conversely, the fact that
information is insensitive to exactly how it is expressed, and can be freely translated from
one form to another, makes it an obvious candidate for a fundamentally important role in
physics, like energy and momentum and other such abstractions. However, until the
second half of this century, the precise mathematical treatment of information, especially
information processing, was undiscovered, so the significance of information in physics
was only hinted at in concepts such as entropy in thermodynamics. It now appears that
information may have a much deeper significance. Historically, much of fundamental
8 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
physics has been concerned with discovering the fundamental particles of nature and the
equations which describe their motions and interactions. It now appears that a different
programme may be equally important: to discover the ways that nature allows, and
prevents, information to be expressed and manipulated, rather than particles to move. For
example, the best way to state exactly what can and cannot travel faster than light is to
identify information as the speed-limited entity. In quantum mechanics, it is highly
significant that the state vector must not contain, whether explicitly or implicitly, more
information than can meaningfully be associated with a given system. Among other
things this produces the wavefunction symmetry requirements which lead to Bose
Einstein and Fermi Dirac statistics, the periodic structure of atoms, and so on.
The programme to re-investigate the fundamental principles of physics from the
standpoint of information theory is still in its infancy. However, it already appears to be
highly fruitful, and it is this ambitious programme that I aim to summarise.
9 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
significant experiments, nuclear magnetic resonance (NMR) has become the most
popular component in quantum hardware architecture. Only within the past year, a group
from Los Alamos National Laboratory and MIT constructed the first experimental
demonstrations of a quantum computer using nuclear magnetic resonance (NMR)
technology. Currently, research is underway to discover methods for battling the
destructive effects of decoherence, to develop an optimal hardware architecture for
designing and building a quantum computer, and to further uncover quantum algorithms
to utilize the immense computing power available in these devices. Naturally this pursuit
is intimately related to quantum error correction codes and quantum algorithms, so a
number of groups are doing simultaneous research in a number of these fields. To date,
designs have involved ion traps, cavity quantum electrodynamics (QED), and NMR.
Though these devices have had mild success in performing interesting experiments, the
technologies each have serious limitations. Ion trap computers are limited in speed by the
vibration frequency of the modes in the trap. NMR devices have an exponential
attenuation of signal to noise as the number of qubits in a system increases. Cavity QED
is slightly more promising; however, it still has only been demonstrated with a few
qubits. Seth Lloyd of MIT is currently a prominent researcher in quantum hardware. The
future of quantum computer hardware architecture is likely to be very different from what
we know today; however, the current research has helped to provide insight as to what
obstacles the future will hold for these devices.
Future Outlook
At present, quantum computers and quantum information technology remains in its
pioneering stage. At this very moment obstacles are being surmounted that will provide
the knowledge needed to thrust quantum computers up to their rightful position as the
fastest computational machines in existence. Error correction has made promising
progress to date, nearing a point now where we may have the tools required to build a
computer robust enough to adequately withstand the effects of decoherence. Quantum
hardware, on the other hand, remains an emerging field, but the work done thus far
suggests that it will only be a matter time before we have devices large enough to test
Shor's and other quantum algorithms. Thereby, quantum computers will emerge as the
superior computational devices at the very least, and perhaps one day make today's
modern computer obsolete. Quantum computation has its origins in highly specialized
fields of theoretical physics, but its future undoubtedly lies in the profound effect it will
have on the lives of all mankind.
BIBILIOGRAPHY:
www.ebi.ac.uk
11 Email: chinna_chetan05@yahoo.com
Visit: www.geocities.com/chinna_chetan05/forfriends.html
References:
1. D. Deutsch, Proc. Roy. Soc. London, Ser. A 400, 97 (1985).
2. R. P. Feynman, Int. J. Theor. Phys. 21, 467 (1982).
3. J. Preskill, "Battling Decoherence: The Fault-Tolerant Quantum
Computer," Physics Today, June (1999).
4. Shor, P. W., Algorithms for quantum computation: Discrete logarithms and
factoring, in Proceedings of the 35th Annual Symposium on Foundations of
Computer Science, IEEE Computer Society Press (1994).
5. Nielsen, M., "Quantum Computing," (unpublished notes) (1999).
6. QUIC on-line, "Decoherence and Error Correction," (1997).
7. D.G. Cory et al., Physical Review Letters, 7 Sept 1998.
8. J. Preskill, "Quantum Computing: Pro and Con," quant-ph/9705032 v3,
26 Aug 1997.
9. Chuang, I. L., Laflamme, R., Yamamoto, Y., "Decoherence and a Simple
Quantum Computer," (1995).
10. D. Deutsch, A. Ekert, "Quantum Computation," Physics World, March
(1998).
Originally Written: 02/25/00
Last Updated: 05/31/00
12 Email: chinna_chetan05@yahoo.com