{"id":2001,"date":"2020-12-04T19:22:51","date_gmt":"2020-12-05T00:22:51","guid":{"rendered":"https:\/\/sites.bu.edu\/phenogeno\/?page_id=2001"},"modified":"2021-03-15T17:56:33","modified_gmt":"2021-03-15T21:56:33","slug":"principal-investigator","status":"publish","type":"page","link":"https:\/\/sites.bu.edu\/phenogeno\/principal-investigator\/","title":{"rendered":"Principal Investigator"},"content":{"rendered":"<h3><img src=\"https:\/\/sites.bu.edu\/phenogeno\/files\/2020\/12\/IMG_7184-769x1024.jpg\" \/><\/h3>\n<h3>My Research Journey (in construction)<\/h3>\n<p>I am sharing these highlights not to brag about accomplishments but to educate and share lessons from many exceptional mentors, students, fellows, collaborators and unusual colleagues I was fortunate to work with over the course of my career.<\/p>\n<h3>Because of my unusual style of <strong><span style=\"color: #ff0000;\">collaboration<\/span><\/strong>, the story also tracks several major developments and historical twists in several fields such as <span style=\"color: #ff0000;\">Artificial Intelligence, Parallel Computing, Machine Learning, Computational Biology\/Bioinformatics, Genomics, Systems Biology, Genomic Network Systems Biology and Medicine, AI applications to Biology and Medicine, Science Informatics, Community Science,<\/span> and more.<\/h3>\n<h3>This review is chronological and spans over 30 years of work. If you are interested in only one topic please skip to it.<\/h3>\n<p>I co-mentored more than 60 students and fellows. I collaborated directly (one on one) with over 70 scientists (not including consortium level collaborators) The full list can be found here:<\/p>\n<h2><a href=\"https:\/\/sites.bu.edu\/phenogeno\/people\/\">My eco-system link<\/a><\/h2>\n<p>&nbsp;<\/p>\n<h3>The main lesson from my long career so far is :<\/h3>\n<h3><span style=\"color: #ffff00;\">Dreaming is easy. Building is hard. Combining dreaming and building is the hardest.<\/span><\/h3>\n<h4><span style=\"color: #ffff00;\">Our eco-system (including lab members and affiliates, collaborators, mentors, resources, and more than anyone in academia, our students or fellows) make &#8220;dreaming and building&#8221; possible. However, we collectively cannot loose track of dreaming. If we just keep up with the present,\u00a0 we would be constantly surprised by the future.\u00a0<\/span><\/h4>\n<h3>Chapter 1:\u00a0 PhD Research (Parallel AI, Parallel Logic Programming (Deductive Databases), Computer Vision)<\/h3>\n<h3>(mentors: Jack Minker and Azriel Rosenfeld)<\/h3>\n<h2><span style=\"color: #ffff00;\">Lesson 1:\u00a0\u00a0<\/span><\/h2>\n<h3><span style=\"color: #ffff00;\"><span style=\"color: #ff0000;\">&#8220;Turf matters as much as seeds&#8221;.<\/span>\u00a0 Many years after PhD I was on an advisory board for Alberta Innovation Academy with several exceptional individuals. One of them was Richard Taylor, a Nobel winner in Physics. Dick said repeatedly &#8220;I grew up on a farm and turf matters as much as seeds&#8221;.\u00a0 <span style=\"color: #ff0000;\">This lesson is often severely underestimated by scientists and college administrators who see science as an individual sport.\u00a0 It rarely is.<\/span><\/span><\/h3>\n<h3><\/h3>\n<h3><span style=\"color: #99cc00;\"><strong>I started my academic journey in Artificial Intelligence.<\/strong><\/span><\/h3>\n<p><strong>I became interested in AI early on mostly because of the work by Zohar Manna on <span style=\"color: #ff0000;\">&#8220;Program Synthesis&#8221;.<\/span>\u00a0 I discovered Zohar Manna&#8217;s work while taking a course from Amir Pnueli.\u00a0 Zohar&#8217;s approach to program synthesis was based on deductive reasoning.\u00a0 Brilliant work but I felt deduction can (and should be) integrated with induction (learning).\u00a0<\/strong><\/p>\n<h3><span style=\"color: #ffff00;\">Originally, I was hoping to <strong>marry computer vision with logic<\/strong> because I thought it would be fundamental and a key to progress in general AI.<\/span><\/h3>\n<p>I brought my own ideas into this mix but the work I did with my PhD co-mentors\u00a0 Azriel and Jack was largely independent of each other with some conceptual leakage but not a real marriage.<\/p>\n<h2><span style=\"color: #ffff00;\"><strong>Lesson 2: <\/strong><\/span><\/h2>\n<h3><span style=\"color: #ffff00;\"><strong>I believe that <span style=\"color: #ff0000;\">integration of logic and perception<\/span> remains a major unsolved AI problem (40 years later).<\/strong><\/span><\/h3>\n<p>I took an unusual path during my PhD because I was co-mentored by two legendary AI researchers who were very different ideologically, philosophically and methodologically. <span style=\"color: #ff0000;\"><strong>Jack Minker<\/strong> <\/span>was following the logic foundation of AI and was focusing on <strong>Logic Programming<\/strong>. He was one of the fathers of <strong>Deductive Databases or what is sometimes called Datalog.\u00a0 <span style=\"color: #ff0000;\">Azriel Rosenfeld <\/span>was one of the fathers of Computer Vision focusing on Perception, Cognition, Computer Vision, Constraints, Neuronal and Pyramidal Networks. <span style=\"color: #ff0000;\">However they shared one important thing in common: a BIG heart and empathy at a rare scale.\u00a0<\/span><\/strong><\/p>\n<h3><strong>I learned many things from Azriel and Jack but the most important lessons were:<\/strong><\/h3>\n<h2><span style=\"color: #ffff00;\">Lessons 3 &amp; 4<\/span><\/h2>\n<ol>\n<li>\n<h3><span style=\"color: #ff0000;\">Have an open mind.<\/span> There is more than one way to study AI (in contrast to several old AI &#8220;gurus&#8221; that treated AI as a religion rather than an expansive field similar to Biology and Physics.\u00a0 \u00a0Open AI should mean <span style=\"color: #ff0000;\">&#8220;Open minded AI&#8221;<\/span> and be inclusive.<\/h3>\n<\/li>\n<li>\n<h3>&#8220;Educating the <span style=\"color: #ff0000;\">mind<\/span> without educating the <span style=\"color: #ff0000;\">heart<\/span> is no education at all&#8221;.. Aristotle but both of my PhD mentors were subscribed to it. They routinely engaged in humanitarian and other activities outside their immediate research area.<\/h3>\n<\/li>\n<\/ol>\n<h3>I co-authored <strong>11+ papers<\/strong> during my PhD. All credit goes to the unusual and creative environment that Azriel and Jack had in their centers.<\/h3>\n<h3><\/h3>\n<h2><span style=\"color: #ffff00;\">Lesson 5:<\/span><\/h2>\n<h3><span style=\"color: #ffff00;\">For some reason that I still cannot fully understand, the <span style=\"color: #ff0000;\">connectionists \/ constraint network \/ semantic network<\/span> parts of the AI world firmly believed in parallelism (Azriel Rosenfeld was definitely a supporter of <span style=\"color: #ff0000;\">parallelism<\/span>).\u00a0 However, the <span style=\"color: #ff0000;\">main stream logic and rule based AI<\/span> scientists had little interest in it. The main lesson to young scientists. Don&#8217;t judge, just follow your passion. Our work with Jack Minker and his group included the <strong>first Parallel Logic Programming System that was actually implemented and tested on a real parallel processor (ZMOB, a 128 processors machine invented by Chuck Rieger).<\/strong> Chuck was an AI visionary who <span style=\"color: #ff0000;\">shared our belief that parallel computing is an important and perhaps even essential trait of AI systems.<\/span> This vision was later reignited by <span style=\"color: #ff0000;\">Marvin Minsky<\/span> and <span style=\"color: #ff0000;\">Danny Hillis<\/span> by building the Connection Machine (a 64000+ processor machine).<\/span><\/h3>\n<p>This was an extremely complex implementation since the machine had NO operating system and we had to implement all messaging and parallel execution on our own.\u00a0 Many highly capable people were involved in this project but much credit goes to Madhur Kohli (a very gifted systems researcher)\u00a0 for carrying the heaviest programming on his shoulders and Jack of course for his early vision.<\/p>\n<h3>We pioneered the first implementation of <span style=\"color: #ff0000;\"><strong>fork join<\/strong> <\/span>in parallel AI. <span style=\"color: #ff0000;\"><strong>Fork-join<\/strong> <\/span>is an fundamental primitive in parallel programming. It was\u00a0 extended significantly many years later to the very famous and widely used <span style=\"color: #ff0000;\"><strong>MAP-REDUCE<\/strong><\/span> paradigm.<\/h3>\n<ol>\n<li>Eisinger, N., S. Kasif and J. Minker, \u201cLogic Programming: A Parallel Approach\u201d, First International Logic Programming Conf., Faculte des Sciences de Luminy Marseille, France, pp.71&#8211;77, September 1982.<\/li>\n<li>Kasif, S., M. Kohli and J. Minker, \u201cPRISM&#8212;A Parallel Inference System Based on Logic\u201d, Logic Programming Workshop, pp.123&#8211;152, Portugal, June 1983.<\/li>\n<li>*Kasif, S., M. Kohli and J. Minker, \u201cControl Facilities of PRISM&#8212;A Parallel Inference System Based on Logic\u201d, International Joint Conf. on Artificial Intelligence, August 1983. <span style=\"color: #ff0000;\"><strong>(Introducing Fork-Join in Parallel Logic Programming).<\/strong><\/span><\/li>\n<li>Chakravarthy, U. S., S. Kasif, M. Kohli, J. Minker and D. Cao, \u201cLogic Programming on ZMOB: A Highly Parallel Machine\u201d, Proc. 1982 International Conf. on Parallel Processing, IEEE Press, pp.347\u2014349 New York, 1982.<\/li>\n<li>C. Asper, D. Cao, U.S. Chakravarthy, A. Csoek-Poeskh, S. Kasif, M. Kohli, J. Minker, R. Piazza and D. Wang, Parallel problem solving on ZMOB,Proc. Trends and Applications 83 (1983) pp. 142\u2013146.<\/li>\n<li>*Kasif, S. and J. Minker, \u201cThe Intelligent Channel: A Scheme for Result Sharing in Parallel Logic Programs\u201d, International Joint Conf. on Artificial Intelligence, pp.29-31, August 1985.<\/li>\n<li>Kasif, S. and A. Rosenfeld, \u201cPyramid Linking as a Special Case of Isodata\u201d, IEEE Transactions on Systems, Man and Cybernetics, Vol.SMC-13, No.1, January 1983. <span style=\"color: #ff0000;\"><strong>(An early approach for Deep Learning (Unsupervised) in Pyramidal Networks used for Image Segmentation. We proved the unsupervised learning algorithm converges).<\/strong><\/span><\/li>\n<li>*Kasif, S. and A. Rosenfeld, \u201cThe Fixpoints of Images and Scenes\u201d, Conf. on Computer Vision and Pattern Recognition, pp.454-456, June 1983. <span style=\"color: #ff0000;\"><strong>(A very abstract formulation of cognition that proposes a mathematical foundation of image\/scene interpretation using\u00a0 monotone operators in Banach Spaces).<\/strong><\/span><\/li>\n<li>Kasif, S., L. Kitchen and A. Rosenfeld, \u201cA Hough Transform Technique for Subgraph Isomorphism\u201d, Pattern Recognition Letters. Vol.2, pp.83&#8211;88, December 1983. <strong>(<span style=\"color: #ff0000;\">An early proposal for matching images and graphs using graph walks. We cannot take any credit for the follow-ups of course but today random walks and their spectral network formulations are widely used for graph and network matching).\u00a0<\/span><\/strong><\/li>\n<li>Simon Kasif, Parallel Searching and Merging on ZMOB. 1984. University of Maryland TR. (Fast parallel searching and merging to implement joins and other queries in deductive databases).<\/li>\n<li>Kasif, S., \u201cControl and Data Driven Execution of Logic Programs: A Comparison\u201d, Journal of Parallel Programming.Vol.15, No.1, pp.73&#8211;100, February 1987.\u00a0 (thesis chapter <strong><span style=\"color: #ff0000;\">complexity analysis of fork join control vs data flow in parallel logic programs, it took a while to publish<\/span>)<\/strong>.<\/li>\n<\/ol>\n<p>&nbsp;<\/p>\n<h3>Fast Chapter 2: Academic Jobs Search<\/h3>\n<p>In 1985 AI was experiencing explosive growth.\u00a0 The timing was fortunate for anyone in AI seeking academic positions.<\/p>\n<p>My area of parallel AI was particularly hot in part due to the <span style=\"color: #ffff00;\"><strong>5th Generation AI Machines Project<\/strong> <\/span>in Japan that forced the US to respond.<\/p>\n<p>But I expect that the positive response to my applications was in part driven by the generosity of my letter writers and their considerable stature. <a href=\"https:\/\/www.cs.hmc.edu\/~fleck\/parable.html\">This academic joke is humorous and self deprecating version of faculty interviews.\u00a0 There is a lot of good work out there of course but the joke is still amusing and provides a perspective.<\/a><\/p>\n<p>I visited eight universities and received seven offers straight out of PhD.<\/p>\n<p>I had a very unusual approach to job applications (not recommended).\u00a0 We intersected nice and relatively small cities ranked high for quality of life in the Almanac with top 20 CS departments. I did not apply to any universities in major cities or places that were not ranked high in quality of life in the Almanac. We also excluded the West Coast.\u00a0 At the completion of the interviews I was mostly considering U. Wisconsin (Madison) and Duke U. But at the end, for complex personal and geographic reasons to stay in the DC area,\u00a0 I applied late to Johns Hopkins University (which was neither top 20 in CS nor in highly ranked city in the Almanac) and joined the department of EECS a week or so after my interview.\u00a0 Johns Hopkins U. is a unique institution and I met many life long friends there. I was particularly influenced by Terry Sejnowski, Fred Jelinek, Vernon Mountcastle, Ham Smith, Adi Karni (I audited his Decision Theory Class as a faculty guest), Sol Snyder (I audited his Neuroscience class as a faculty guest) and many direct collaborators and students.<\/p>\n<p>As mentioned,\u00a0 our academic generation was very lucky.\u00a0 Five years later the market dried up, AI was experiencing a winter and it became very challenging to obtain an academic position in AI.<\/p>\n<p>In 2020 (today) AI applicants are lucky again as deep learning is driving enormous interest and hiring in AI.<\/p>\n<p>&nbsp;<\/p>\n<h2><span style=\"color: #ffff00;\">Lesson 6:\u00a0<\/span><\/h2>\n<h3><span style=\"color: #ffff00;\"><strong>I am describing this experience in detail because many young PhDs take their interviews way too personally. The faculty interview success is heavily dependent on many factors that are completely independent of the candidate&#8217;s ability and accomplishments.\u00a0 So relax and enjoy the ride.\u00a0 Five years later it may change completely :)<\/strong><\/span><\/h3>\n<p>&nbsp;<\/p>\n<h3>Chapter 3: Early Academic Research (Parallel AI)<\/h3>\n<h3>(with Art Delcher and Lewis Stiller)<\/h3>\n<p>After joining Hopkins,\u00a0 I din&#8217;t want to continue with heavy software development work and implementations of parallel AI systems. I realized that parallel machines were not ready for us and the overhead of developing the operating system, messaging, and parallel execution on raw architectures were exceptionally challenging but had little todo with basic AI research. Also Hopkins was\u00a0 very small (4 CS faculty total) and unlike universities such as CMU or Illinois the group did not have opportunities for collaboration with computer systems researchers.<\/p>\n<p>I decided to shift research interests and <strong><span style=\"color: #ff0000;\">marry parallel algorithms and complexity with AI.<\/span><\/strong> Together with exceptional PhD students we produced several foundational results developing basic AI algorithms on abstract parallel models of computation.\u00a0 It was a niche area since parallel algorithms were largely studied by CS theorists who focused on basic graph problems or sorting not AI algorithms.\u00a0 Most of core AI people at that time worked on heuristic search or ad-hoc machine learning ignoring parallel machines (with few notable exceptions).<\/p>\n<p>This\u00a0 allowed our group to have relatively little competition in this space and we established many basic theoretical and other textbook results that stood the test of time. Selected results are listed below and the surprising or fundamental results are highlighted. All the work was theoretical (except the amazing work by Lewis Stiller on the Connection Machine 2).<\/p>\n<ol>\n<li><strong>Optimal Parallel Term Matching (with Art Delcher) on Shared Memory Parallel Machines (textbook results)<\/strong><\/li>\n<li><strong style=\"color: #ffff00;\">2. Lower Bounds on Parallel Constraint Satisfaction (on my own) &#8211; A fundamental AI constraint propagation procedures previously believed to be highly parallelizable is inherently sequential (textbook result in AI).<\/strong><\/li>\n<li><strong>Optimal Parallel Anti-Unification (with Art Delcher) on Shared Memory Parallel Machines (textbook results)<\/strong><\/li>\n<li>Parallel Propositional Satisfiability (with John Reif and Deepak Sherlekar)<\/li>\n<li><span style=\"color: #ffff00;\"><strong>Parallel log time Inference in Bayes Tree Networks (with Art Delcher)<\/strong><\/span><\/li>\n<li><strong style=\"color: #ffff00;\"><a href=\"https:\/\/www.nytimes.com\/1991\/10\/30\/us\/computer-is-pushed-to-edge-to-solve-old-chess-problem.html\" style=\"color: #ffff00;\">Parallel Chess End Game Analysis (Lewis Stiller PhD Thesis) &#8211; a classical result in Computer Chess<\/a><\/strong><\/li>\n<li><strong>Incremental Parallel Computation (with Art Delcher, a new complexity class highlighting problems that are difficult to solve incrementally)<\/strong><\/li>\n<li><strong><span style=\"color: #ffff00;\">Parallel Inference in Bayes Networks (with Art Delcher, Adam Grove and Judea Pearl<\/span><\/strong><\/li>\n<\/ol>\n<p>I was very fortunate working with exceptional PhD students in Parallel AI who walked on water. In particular, <span style=\"color: #ff0000;\"><strong>Art Delcher<\/strong><\/span> became a life long friend and collaborator. <strong><span style=\"color: #ff0000;\">Lewis Stiller<\/span><\/strong> did exceptionally challenging research on chess end game analysis on a 65000 processor parallel machine (Thinking Machines, CM2). He used techniques from computer algebra and group theory integrated with AI and parallel software development.<\/p>\n<p>We also had a number of very gifted undergraduate students rotating with us and doing projects in parallel AI. They are listed on the eco system page.<\/p>\n<h2><span style=\"color: #ffff00;\">Lesson 7:\u00a0<\/span><\/h2>\n<h3><span style=\"color: #ffff00;\">We learned many profound theoretical, technical and practical ideas, frameworks, and methods.\u00a0 However, one particular lesson that stands out is a question posed by then undergraduate research student Phoebe Sengers who was programming on the Connection Machine (64000+ processors) as an undergraduate project.\u00a0 <span style=\"color: #ff0000;\">&#8220;What would AI look like if it was pioneered and driven by women?&#8221;<\/span>\u00a0 The most influential people historically early on were Marvin Minsky, John McCarthy, Allen Newell and Herb Simon.\u00a0 Phoebe is now a professor at Cornell working at the intersection of culture and Science .<\/span><\/h3>\n<p>&nbsp;<\/p>\n<h3>Chapter 4:\u00a0 Machine Learning: Theory and Widely Used System Development<\/h3>\n<h3>(with Steven Salzberg, David Heath, S. Murthy, John Rachlin, David Waltz and others)<\/h3>\n<p>I wanted to go back to software development and because of Lewis Stiller&#8217;s work read up on Decision Diagrams and Decision Trees. At the same time Steven Salzberg joined Hopkins. Steven was an applied machine learning researcher who was interested in developing and validating ML software. Our collaboration proved very fruitful because I was interested in theoretical and conceptually novel developments and Steven cared deeply about dissemination, testing and validation.<\/p>\n<p>We collaborated both on systems development and novel theoretical ideas.<\/p>\n<p>The main topics are highlighted below.<\/p>\n<ol>\n<li>We introduced a novel model for learning (data mining) using limited memory (1991). This development proceeded the broad interest in data-streaming that became a major field in theoretical CS and data mining. We proved both upper and lower bounds on the number of rounds needed for mining data.\u00a0 We also suggested a cognitive motivation.<\/li>\n<li>We pioneered a new model for Learning with a Helpful Teacher that allows to learn from very few examples.<\/li>\n<li>We developed a widely used open access system for decision tree induction, OC1. This system introduced a number of new ideas in decision tree induction:\n<ol>\n<li>Sorting on attributes to enable scalable learning<\/li>\n<li>Randomization (prior to Random Forest).\u00a0 For the record Random Forest randomization was more rigorously conceived and analyzed.<\/li>\n<li>\u00a0We implemented a voting approach on multiple decision trees using randomization (also prior to Random Forest).<\/li>\n<\/ol>\n<\/li>\n<li>With David Waltz we proposed the first Kernel based on Bayes Networks (BM) to implement a K-NN learning approach using the\u00a0 BN transformed data.<\/li>\n<\/ol>\n<h2><span style=\"color: #ffff00;\">Lesson 8: <\/span><\/h2>\n<h3><span style=\"color: #ffff00;\"><span style=\"color: #ff0000;\">In my view we do not have enough long term and sustainable direct collaborations in AI between applied and theoretical research.<\/span> However, a review of the high impact literature suggests many successful examples (e.g., McCallum\/Lafferty\/Pereira, Blum\/Mitchell, Freund\/Shapire, LISP\u00a0 and many, many more).\u00a0 <span style=\"color: #ff0000;\">Many funded projects do lip service to these theory-practice collaborations but largely leave the participants in their own caves.\u00a0<\/span> It takes special people without fragile egos and ability to mutually appreciate the intellectual challenge in doing both theory and experimental research at the highest level.\u00a0 At the lowest level, theory is indeed a superior discipline at the intellectual level because it requires formal proofs at any level of the field. <span style=\"color: #ff0000;\">However, at the highest level experimental research is as magical and elegant as any theory.\u00a0 I personally remain fully committed to both.\u00a0<\/span> However, one needs to be able to selectively lower the bar on either theoretical or empirical methodologies to identify novel paths at the intersection of both. This is one of the great challenges for scientists in these philosophically and methodologically different areas.\u00a0 However, if we learn how to simultaneously advance theory and experiments, <span style=\"color: #ff0000;\">the effective translation to practice can be shortened significantly with tremendous benefits to society.\u00a0 Which is what matters the most!\u00a0<\/span><\/span><\/h3>\n<p>&nbsp;<\/p>\n<h2>Chapter 5:\u00a0 Bayes Networks (Graphical Models) and Biology<\/h2>\n<h3>We helped conceive and help popularize the use of Bayes networks (Graphical Models) in Biology starting 1992.<\/h3>\n<h3><span style=\"color: #ff0000;\">We made two predictions\u00a0 in our 1993-5 papers that provide a good motivation for the utility of graphical models in reasoning about biology or modeling biological systems.<\/span><\/h3>\n<h3>&#8220;To summarize, scientific\f analysis of data is an important potential application of Artificial Intelligence (AI) research. We believe that the ultimate data analysis system using AI techniques will have a wide range of tools at its disposal and will adaptively choose various methods.<\/h3>\n<h3>It will be able to generate simulations automatically and verify the model it constructed with the data generated during these simulations. When the model does not fi\ft the observed results the system will try to explain the source of error, conduct additional experiments, and choose a different model by modifying system parameters. If it needs user assistance, it will produce a simple low-dimensional view of the constructed model and the data. This will allow the user to guide the system toward constructing a new model and\/or generating the next set of experiments.&#8221;<\/h3>\n<h2><span style=\"color: #ffff00;\">Prediction 1: <\/span><\/h2>\n<h2><span style=\"color: #ffff00;\">&#8220;We believe that flexibility, efficiency and direct representation of causality in probabilistic networks are important desirable features that make them very strong candidates as a framework for biological modeling systems.&#8221;<\/span><\/h2>\n<h4><span style=\"color: #ff0000;\">This quote From Delcher, A., S. Kasif, H. Goldberg and W. Xsu, \u201cProtein Secondary-Structure Modeling with Probabilistic Networks\u201d, International Conference on Intelligent Systems and Molecular Biology, pp. 109\u2013117, 1993.<br \/>\n<\/span><\/h4>\n<h4><span style=\"color: #ff0000;\">Delcher, A., S. Kasif, H. Goldberg and W. Xsu, \u201cApplication of Probabilistic Causal Trees to Analysis of Protein Secondary Structure\u201d, Proceedings of the National Conference on Artificial Intelligence, pp. 316\u2013321, July 1993.<\/span><\/h4>\n<h2><span style=\"color: #ffff00;\">Prediction 2: <\/span><\/h2>\n<h3><span style=\"color: #ffff00;\">Based on Prediction 1, we also proposed a relatively efficient computational framework to implement in-silico directed evolution (synthetic biology) using graphical models.\u00a0 More specifically, the framework must perform mutagenesis\u00a0 (e.g., amino acid substitutions in proteins) and screen for mutations and adaptations that satisfy structural and functional constraints.\u00a0 For each such perturbation (mutation) we must perform an efficient inference in the graphical model to assess the probability a given property of a protein or system was affected by the perturbation. Our novel procedure enables to <span style=\"color: #ff0000;\">COMPILE THE GRAPHICAL MODEL<\/span> using dynamic data structures and subsequently compute the change in a specific property in time that is exponentially faster than using normal algorithms.\u00a0 \u00a0This is described in detail in the paper below.\u00a0<\/span><\/h3>\n<p>&nbsp;<\/p>\n<h3><span style=\"color: #ff0000;\">Delcher, A, A. Grove, S. Kasif and J. Pearl, <a href=\"https:\/\/www.jair.org\/index.php\/jair\/article\/view\/10154\/24067\" style=\"color: #ff0000;\">\u201cLogarithmic Time Queries and Updates in Probabilistic Networks\u201d<\/a>, Journal of Artificial Intelligence Research, Vol. 4., pp. 37\u201359, 1996<\/span><\/h3>\n<p>&nbsp;<\/p>\n<h3>Chapter 6. Computational Genomics (Human Genome Project, Before and After)<\/h3>\n<h4>(With Steven Salzberg, Art Delcher, Owen White, Herve Tettelin, many amazing TIGR biologists, Human Genome Project Consortium and many pioneers in computational genomics, Rich Roberts, Charles Cantor, Yu Zheng, Megon Walker, Chunmin Ding,\u00a0 Charles DeLisi, Zhiping Weng and many, many more.)<\/h4>\n<h4><span style=\"color: #ff0000;\">Selected Results, Ideas, Frameworks:\u00a0<\/span><\/h4>\n<h4><a href=\"https:\/\/en.wikipedia.org\/wiki\/GLIMMER\">GLIMMER (one of the most widely used gene finders in bacterial genomes)<\/a><\/h4>\n<h4><a href=\"https:\/\/en.wikipedia.org\/wiki\/MUMmer\">MUMMER (widely used comparative genome system, first open access whole genome comparison system for bacteria).<\/a><\/h4>\n<h3>PUBLIC HUMAN GENOME PROJECT (HGP) &#8211; with an amazing group of collaborators that include many computational biology pioneers<\/h3>\n<h4><a href=\"https:\/\/pubmed.ncbi.nlm.nih.gov\/10842737\/\">BAYES NETS FOR ANALYSIS OF SPLICING (with Art Delcher)<\/a><\/h4>\n<h3><a href=\"https:\/\/academic.oup.com\/bioinformatics\/article\/18\/1\/19\/243521\">Bayes Networks for Genomic Data Integration and Genome Annotation<\/a><\/h3>\n<h4><a href=\"https:\/\/www.hpl.hp.com\/techreports\/Compaq-DEC\/CRL-2001-5.pdf\">REMOTE HOMOLOGY PREDICTION USING SIMPLE KERNELS FOR Human Genome Annotation<\/a> (with Pedro Moreno and others at HP Labs, summer 2000)<\/h4>\n<h3><a href=\"https:\/\/www.researchgate.net\/publication\/7762401_MuPlex_Multi-objective_multiplex_PCR_assay_design\">MULTIPLEX PCR (for Prenatal Diagnostics of Down Syndrome and Liquid Biopsy).<\/a><\/h3>\n<h3>(With Charles Cantor, John Rachlin, Noga Alon and others.)<\/h3>\n<h3><a href=\"https:\/\/pubmed.ncbi.nlm.nih.gov\/12176930\/\">Early Network and Genomic Data Integration for Genome Annotation and Gene Function Prediction<\/a><\/h3>\n<h3>(with Rich Roberts and Yu Zheng)<\/h3>\n<h3>COMPARATIVE ANALYSIS (many papers in human, mouse, model organisms, bacterial genomes).<\/h3>\n<h4>and more &#8230;<\/h4>\n<h2><span style=\"color: #ffff00;\">Lesson 10: <\/span><\/h2>\n<h3><span style=\"color: #ffff00;\">We were very lucky to be part of this transformative period but there is a quote that captures this better than most.<\/span><\/h3>\n<h3><span style=\"color: #ffff00;\">&#8220;Fortune favors the prepared mind&#8217;, Louis Pasteur<\/span><\/h3>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<h3><strong>Chapter 7. Computational Genomic Network Systems Biology <\/strong><\/h3>\n<h3>(with Rich Roberts, Yu Zheng, Stan Letovsky and many others)<\/h3>\n<h3><span style=\"color: #ff0000;\">1. Network Based Function Prediction (2002-)<\/span><\/h3>\n<h4>Partners: Stan Letovsky, T.M. Murali, Naoki Naryai, Charles Cantor, Rich Roberts,\u00a0 and many others.<\/h4>\n<h3>We helped conceive and popularize systematic whole genome network based function prediction.\u00a0 Our work was derived from frameworks previously used in AI, Computer Vision and CS that evolved from Markov Random Fields,\u00a0 Neural Nets (Hopfield Networks), Graph Cuts, Graph Diffusion and Label Propagation in Networks. Today, network based function prediction is a widely used framework for the field.<\/h3>\n<p>&nbsp;<\/p>\n<h3>2. Network Based Disease Biomarkers for Diabetes and Metabolic Disease<\/h3>\n<h4>Partners: Zak Kohane, Ron Kahn, Manway Liu, Terrence Wu, Tianxi Cai and others as part of the I2B2 NIH National Center at Harvard Partners)<\/h4>\n<h3><span style=\"color: #ff0000;\">3. Multiplex Networks,\u00a0 Biological Context Networks, Multi-node graphs (2004-)<\/span><\/h3>\n<h4>(With Noga Alon, Vera Asodi, Charles Cantor, John Rachlin and others)<\/h4>\n<p>&nbsp;<\/p>\n<h3>4. Regulatory Networks\u00a0 Discovery and Experimental Validation (2001-)<\/h3>\n<h4>(With Geoff Cooper, John Tullai, Michael Schaffer, Jim Collins, Tim Gardner, J. Faith, Boris Hayette, Esther Rheinbay, Mario Suva, Brad Bernstein and many more.)<\/h4>\n<p>&nbsp;<\/p>\n<h3><span style=\"color: #ff0000;\">Chapter 8. Rich Roberts and the Computational Bridges to Experiments (COMBREX) Project<\/span><\/h3>\n<p>&nbsp;<\/p>\n<h3>Chapter 9.\u00a0 I2B2 (Informatics to Bedside) at Harvard Partners<\/h3>\n<h3>(with Zak Kohane, Tianxi Cai,\u00a0 ME Patti, Ron Kahn, Terrence Wu and many others).<\/h3>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<h3>Chapter 10. Joslin Diabetes NIH Center and Regional Systems Biology Core (Diabetes Type 2, Aging)<\/h3>\n<h3>(With Michael Molla, Jonathan Dreyfuss, Ron Kahn, ME Patti, Allison Goldfine, George King and others)<\/h3>\n<p>&nbsp;<\/p>\n<h3><span style=\"color: #ffff00;\">Lesson 11:<\/span><\/h3>\n<h3><span style=\"color: #ffff00;\">Based on my experience working with Joslin\u00a0 and inspired by the quote &#8220;Make things as simple as possible\u00a0 but not any simpler&#8221;, it is tempting to develop a <span style=\"color: #ff0000;\">theory of wellness which is minimal, supported by evolution and enables direct control.\u00a0<\/span><\/span><\/h3>\n<p>&nbsp;<\/p>\n<h2><span style=\"color: #ffff00;\">Chapter 11. AI2BIO<\/span><\/h2>\n<p>&nbsp;<\/p>\n<h2><span style=\"color: #ffff00;\">Lesson : <\/span><\/h2>\n<h3><span style=\"color: #ffff00;\">After working for almost 30 years at the intersection of AI and Biology I believe that the integration of AI and Biology has unlimited potential. This goes way beyond the old arguments about biology inspired AI (e.g., evolutionary algorithms, neural computation, etc).\u00a0 I am alluding to the transformative methodologies developed in AI that can be used to understand nature.\u00a0 In fact, in a recent talk at MIT I have argued <a href=\"https:\/\/www.csail.mit.edu\/event\/should-machines-understand-nature-pass-turing-test-co-evolving-ai-and-systemssynthetic\">&#8220;Should Machines Understand Nature to Pass the Turing Test? Co-evolving AI and Systems\/Synthetic Biology&#8221;<\/a><\/span><\/h3>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<h3>AI and Biology (with Rich Roberts)<\/h3>\n<p>&nbsp;<\/p>\n<h3><a href=\"https:\/\/arxiv.org\/abs\/2010.12015\">AI and Ethics (review)<\/a><\/h3>\n<p>&nbsp;<\/p>\n<h3><a href=\"https:\/\/journals.plos.org\/plosbiology\/article?id=10.1371\/journal.pbio.3000999\">Tracing Predictions to Experimental Ground Truth (With Rich Roberts)<\/a><\/h3>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<h3>Chapter 12: ?<\/h3>\n<p>&nbsp;<\/p>\n<h3><\/h3>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n","protected":false},"excerpt":{"rendered":"<p>My Research Journey (in construction) I am sharing these highlights not to brag about accomplishments but to educate and share lessons from many exceptional mentors, students, fellows, collaborators and unusual colleagues I was fortunate to work with over the course of my career. Because of my unusual style of collaboration, the story also tracks several [&hellip;]<\/p>\n","protected":false},"author":8782,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":[],"_links":{"self":[{"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/pages\/2001"}],"collection":[{"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/users\/8782"}],"replies":[{"embeddable":true,"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/comments?post=2001"}],"version-history":[{"count":50,"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/pages\/2001\/revisions"}],"predecessor-version":[{"id":2579,"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/pages\/2001\/revisions\/2579"}],"wp:attachment":[{"href":"https:\/\/sites.bu.edu\/phenogeno\/wp-json\/wp\/v2\/media?parent=2001"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}