Robert Coop

Robert Coop

Nashville Metropolitan Area
3K followers 500+ connections

About

Dr. Robert Coop, also known as Coop, has over 15 years of experience in the data science…

Activity

Join now to see all activity

Experience

  • Lapis Legal AI Graphic

    Lapis Legal AI

    Nashville, Tennessee, United States

  • -

    Nashville, TN

  • -

  • -

  • -

    Nashville, Tennessee, United States

  • -

  • -

  • -

  • -

    Greater Atlanta Area

  • -

    Greater Atlanta Area

  • -

    Greater Atlanta Area

  • -

    Greater Atlanta Area

  • -

    Greater Atlanta Area

  • -

    Knoxville, Tennessee

  • -

    Sierra Vista, Arizona

  • -

  • -

  • -

    Knoxville, Tennessee Area

  • -

  • -

    Knoxville, Tennessee Area

Education

  • University of Tennessee, Knoxville Graphic

    University of Tennessee-Knoxville

    -

    -

    Activities and Societies: Machine Intelligence Lab

    Studying machine learning and artificial intelligence, with a focus in neural networks, genetic algorithms, and reinforcement learning.

  • -

    -

  • -

    -

Licenses & Certifications

Publications

  • Mitigation of catastrophic interference in neural networks using a fixed expansion layer

    Circuits and Systems (MWSCAS), 2012 IEEE 55th International Midwest Symposium on

    In this paper we present the fixed expansion layer (FEL) feedforward neural network designed for balancing plasticity and stability in the presence of non-stationary inputs. Catastrophic interference (or catastrophic forgetting) refers to the drastic loss of previously learned information when a neural network is trained on new or different information. The goal of the FEL network is to reduce the effect of catastrophic interference by augmenting a multilayer perceptron with a layer of sparse…

    In this paper we present the fixed expansion layer (FEL) feedforward neural network designed for balancing plasticity and stability in the presence of non-stationary inputs. Catastrophic interference (or catastrophic forgetting) refers to the drastic loss of previously learned information when a neural network is trained on new or different information. The goal of the FEL network is to reduce the effect of catastrophic interference by augmenting a multilayer perceptron with a layer of sparse neurons with binary activations. We compare the FEL network's performance to that of other algorithms designed to combat the effects of catastrophic interference and demonstrate that the FEL network is able to retain information for significantly longer periods of time with substantially lower computational requirements.

    Other authors
    See publication
  • Mapping the Landscape of Human-Level Artificial General Intelligence

    AI Magazine, 33(1)

    We present the broad outlines of a roadmap toward humanlevel artificial general intelligence (henceforth, AGI). We begin by discussing AGI in general, adopting a pragmatic goal for its attainment and a necessary foundation of characteristics and requirements. An initial capability landscape will be presented, drawing on major themes from developmental psychology and illuminated by mathematical, physiological and information processing perspectives. The challenge of identifying appropriate tasks…

    We present the broad outlines of a roadmap toward humanlevel artificial general intelligence (henceforth, AGI). We begin by discussing AGI in general, adopting a pragmatic goal for its attainment and a necessary foundation of characteristics and requirements. An initial capability landscape will be presented, drawing on major themes from developmental psychology and illuminated by mathematical, physiological and information processing perspectives. The challenge of identifying appropriate tasks and environments for measuring AGI will be addressed, and seven scenarios will be presented as milestones suggesting a roadmap across the AGI landscape along with directions for future research and collaboration.

    Other authors
    See publication
  • Functional Analysis and Prediction of Brain Tumor Growth

    BSEC 2011: 3rd Annual ORNL Biomedical Science and Engineering Conference

    This work introduces a method for predicting tumor growth by:
    1) Identifying functional masks for the tumor with high predictive power
    2) Learning parameters that predict growth based on these masks
    Given simulated tumor growth trajectories (presented as cancer cell counts for voxels within a 128x128 grid), we first examine a large number of trajectories in order to determine, for any given central voxel, which neighboring voxels have the
    highest predictive power in…

    This work introduces a method for predicting tumor growth by:
    1) Identifying functional masks for the tumor with high predictive power
    2) Learning parameters that predict growth based on these masks
    Given simulated tumor growth trajectories (presented as cancer cell counts for voxels within a 128x128 grid), we first examine a large number of trajectories in order to determine, for any given central voxel, which neighboring voxels have the
    highest predictive power in determinining future states of the central voxel (we refer to these sets of predictive voxels as 'functional masks'). A novel functional analysis procedure is developed which uses an evolutionary algorithm combined with general systems theory principles in order to search for highly predictive functional masks. Probabilistic parameter estimation is then performed over the voxels in this
    functional mask in order to determine the probability distribution of the next state of the central voxel conditioned on the current and previous states of its neighbors.

    Using this approach we are able to make accurate predictions of tumor growth using general methods that do not rely on biological or other domain­ specific knowledge about the detailed mechanisms underlying the growth.

    See publication
  • DeSTIN: A Scalable Deep Learning Architecture with Application to High-Dimensional Robust Pattern Recognition

    Proc. AAAI 2009 Fall Symposium on Biologically Inspired Cognitive Architectures

    The topic of deep learning systems has received significant attention during the past few years, particularly as a biologically-inspired approach to processing high-
    dimensional signals. The latter often involve spatiotemporal information that may span large scales, rendering its representation in the general case highly challenging. Deep learning networks attempt to overcome this challenge by means of a hierarchical architecture that is comprised of common
    circuits with similar (and…

    The topic of deep learning systems has received significant attention during the past few years, particularly as a biologically-inspired approach to processing high-
    dimensional signals. The latter often involve spatiotemporal information that may span large scales, rendering its representation in the general case highly challenging. Deep learning networks attempt to overcome this challenge by means of a hierarchical architecture that is comprised of common
    circuits with similar (and often cortically influenced) functionality. The goal of such systems is to represent sensory
    observations in a manner that will later facilitate robust pattern classification, mimicking a key attribute of the mammal
    brain. This stands in contrast with the mainstream approach of pre-processing the data so as to reduce its dimensionality
    - a paradigm that often results in sub-optimal performance.
    This paper presents a Deep SpatioTemporal Inference Network (DeSTIN)- ascalable deeplearning architecture that relies on a combination of unsupervised learning and Bayesian inference. Dynamic pattern learning forms an inherent way of capturing complex spatiotemporal dependencies. Simulation results demonstrate the core capabilities of the proposed framework, particularly in the context of high-dimensional signal classification.

    Other authors
    See publication

Courses

  • Applied Linear Algebra: Engineering Systems

    529

  • Bayesian Statistics I

    525

  • Data Mining

    594

  • Detection/Estimation Theory

    643

  • Discrete Event Systems

    617

  • Graphical Models in Artificial Intelligence

    692

  • Mathematical Modeling

    411

  • Nonlinear Systems Theory

    613

  • Optimization (Computer Engineering Applications)

    617

  • Optimization (Mathematical Theory)

    577

  • Pattern Recognition

    571

  • Random Process Theory

    504

  • Real Analysis

    545

  • Statistical Methods in Industrial Engineering

    516

Projects

  • Functional Analysis of Cellular Automata

    -

    Project satisfying thesis requirement for a Master of Science degree.

    In this project, generalized 2-dimensional binary cellular automata are modeled using the discrete event system specification (DEVS).

    Cellular automata are randomly generated and simulated. Observations of these simulations are used in combination with an intelligent, entropy based evolutionary algorithm in order to infer the transition function of the cellular automaton.

    See project

Honors & Awards

  • Best Poster – International Joint Conference on Neural Networks

    International Joint Conference on Neural Networks

    Presented my dissertation work (Ensemble Learning in Fixed Expansion Layer Networks for Mitigating Catastrophic Forgetting) on a poster. My poster was selected from among over thirty other posters for this award.

Recommendations received

8 people have recommended Robert

Join now to view

More activity by Robert

View Robert’s full profile

  • See who you know in common
  • Get introduced
  • Contact Robert directly
Join to view full profile

Other similar profiles

Explore top content on LinkedIn

Find curated posts and insights for relevant topics all in one place.

View top content

Add new skills with these courses