21.05.2020 Views

Broad Street Scientific Journal 2020

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

BROAD

STREET

SCIENTIFIC

VOLUME 9 | 2019-20

The North Carolina School of Science and Mathematics Journal of

Student STEM Research


Front Cover

Human Eye

The human eye is one of the most complex parts of

our physiology. Our eyes can differentiate between

approximately 10 million different colors and can even

detect a single photon of light. In the eye, light travels

through the cornea and pupil to the crystalline lens,

which focuses the light onto the retina. The light falling

onto the retina is converted to electrical signals that

travel through the optic nerve to the brain, creating

what we experience as vision. Image Credit: Dhyamis

Kleber.

Approximate Scale: 20 millimeters

Biology Section

Tissue with Bronchogenic Carcinoma

Tissue is an extracellular matrix of similar cells from the

same origin that together carry out a specific function.

Most cancers arise in epithelial tissue. One such example

is Bronchogenic Carcinoma, a lung cancer that begins in

the tissue lining the airway of the lungs and spreads outwards,

creating more and more mutated cells, eventually

forming a tumor. Lung and bronchus cancers account

for about 13 percent of new cancer cases in the United

States. Image Credit: Microscopy U.

Approximate Scale: 250 micrometers

Chemistry Section

Fluorescence Image of Neuronal Culture

The cell is the basis for all life on Earth. From the 30

meter long Blue Whale to the 370 nanometer long Pelagibacter

ubique, every living organism on the planet

consists of at least one cell. Every individual cell consists

of genetic material suspended in a cytoplasm contained

within a membrane. The human body is made of about

37.2 trillion cells. Image Credit: Jan Schmoranzer, 2011

Olympus BioScapes Digital Imaging Competition.

Approximate Scale: 60 micrometers


Engineering Section

TIGER Domain

Organelles are to cells like organs are to the body. They are

smaller than cells and harder to see, so new discoveries are

revolutionary and unique. The TIGER domain, discovered

in 2018, is a space where messenger RNAs (mRNAs)

encoding certain kinds of proteins find the appropriate

environment to grow. TIGER stands for TIS granules

(purple) and Endoplasmic Reticulum (green). The central

black area is the cell nucleus. Image Credit: Dr. Christin

Mayr, Memorial Sloan Kettering Cancer Center.

Approximate Scale: 25 micrometers

Mathematics and Computer Science Section

Side View Image of Live SKOV3 (Ovarian Cancer) cell

Proteins are complex molecules that play critical roles in

the human body and are required for the structure, function,

and regulation of the body’s tissues and organs. Sideview

images of a live SKOV3 (Ovarian Cancer) cell show

it undergoing compression by an an atomic force microscope.

The green shows the fluorescently tagged protein

Histone 2B, while the magenta shows KRas-tail, a protein

that localizes to the cell membrane. Image Credit: Chad M.

Hobson, UNC-Chapel Hill Department of Physics and Astronomy.

Approximate Scale: 10 micrometers

Physics Section

Nine Essential Amino Acids

Amino acids are the organic building blocks of proteins. In

the body, proteins are used to facilitate chemical reactions,

provide structure, act as chemical messengers, store energy,

and transport other molecules. Out of the 20 amino acids

used by humans, 9 are essential amino acids, which cannot

be made by the body and must be supplied by our diet.

These 9 essential amino acids are histidine, isoleucine, leucine,

lysine, methionine, phenylalanine, threonine, tryptophan,

and valine. Image Credit: MolView, Eleanor Xiao.

Approximate Scale: 0.4-1 nanometer per amino acid

Back Cover

Color-coded Distribution of Neurons in Transgenic

Mouse Dentate Gyrus

To understand the brain, scientists need to map how neurons

are wired to form circuitries in healthy and disease

states. The traditional approach of tissue slicing is difficult

and labor intensive as the neuronal circuitries are vast,

three dimensional, and tightly entangled. A new tissue

clearing solution, OPTIClear, selectively adjusts the optical

properties of tissue without damaging their structural

components. Image Credit: Dr. Wutian Wu, The University

of Hong Kong (HKU) and Imperial College London.

Approximate Scale: 500 micrometers


TABLE of CONTENTS

4 Letter from the Chancellor

5 Words from the Editors

6 Broad Street Scientific Staff

7 Essay: The Forsaken Victims of Climate Change

MEGAN MOU, 2021

Biology

10 Small Molecule Activation of Wnt/β-Catenin Signaling Pathway on Neurodegeneration

Rates of Dopaminergic Neurons in C. elegans

ARIBA HUDA, 2020

18 Quorum Quenching: Synergistic Effects of Plant-Derived Compounds on Biofilm

Formation in Vibrio harveyi

PREETI NAGALAMADAKA, 2020

25 Emotional Processing and Working Memory in Schizophrenia with Neutral and

Negative Stimuli: An fMRI Study

CINDY ZHU, 2020

Chemistry

33 Sesamol as a Novel Redox Mediator for the Electrochemical Separation of Carbon

Dioxide from Flue Gas

MADISON HOUCK, 2020

40 Modeling the Effect of Chemically Modified Non-Antibiotic Tetracyclines with

β-Amyloid Fibrils to Treat Alzheimer's Disease

RACHEL QU, 2020

46 Synthesis of a Tau Aggregation Inhibitor Related to Alzheimer's Disease

EMMA STEUDE, 2020


Engineering

51 A Telemetric Gait Analysis Insole and Mobile App to Track Postoperative Fracture

Rehabilitation

SAATHVIK A. BOOMPELLI, 2021 ONLINE

57 Differences in Reliability and Predictability of Harvested Energy from Battery-less

Intertermittently Powered Systems

NITHYA SAMPATH, 2020

Mathematics and Computer Science

62 Applying Machine Learning to Heart Disease Diagnosis: Classification and Correlation

Analyses

SAHIL PONTULA, 2020 ONLINE

69 Modeling the Effect of Stem Cell-Targeted Immunotherapy on Tumor Size

AMBER POSPISTLE, 2021 ONLINE

Physics

76 The Cosmic Radiation Shielding Properties of Lunar Regolith

ELEANOR MURRAY, 2020

81 Modeling the Gravitational Lens of the Einstein Ring MG1131+0456

NATHANIEL S. WOODWARD, 2020

Featured Article

89 An Interview with Mr. Erik Troan


LETTER from the CHANCELLOR

"For 300 days I’ve been fortunate to be a part of something that strives to represent and benefit all humanity. As I

look out and reflect on our shared world, that has been the greatest honor. . . I can’t wait to pay it forward to the next

explorers and watch them fly even higher."

~ Christina H. Koch ‘97

I am very happy to introduce the ninth edition of

the North Carolina School of Science and Mathematics’

(NCSSM) scientific journal, Broad Street Scientific, with the

words of NASA astronaut and NCSSM graduate Christina

H. Koch ‘97, who recently set a record for the longest single

spaceflight by a woman. We are proud of Christina, and

of each of our NCSSM students who conduct significant

scientific research during the academic year. Broad Street

Scientific is a student-led and student-produced showcase

of some of the impressive research currently being

conducted by students at NCSSM.

As I write this foreword, those of us in North

Carolina and in much of the United States are responding

to stay-at-home orders as a result of the COVID-19 global

pandemic. The importance of scientific inquiry is amplified

in times like these, when cutting-edge technologies and

treatments are literally saving lives each day, and our

researchers accelerate their search for both a cure for

this virus and a vaccine to prevent its return. At NCSSM,

students are given regular opportunities to develop their

research skills and share their results as part of the broader

scientific conversation around the world. Many go on to

join that larger community of researchers, asking difficult

questions and providing guidance grounded in data to

those who may address the next global challenge. We are

proud to be a part of their educational journey.

Opened in 1980, NCSSM was the nation’s first public

residential high school where students study a specialized

curriculum emphasizing science and mathematics.

Teaching students to do research and providing them with

opportunities to conduct high-level research in biology,

chemistry, physics, computational science, engineering

and computer science, math, humanities, and the social

sciences is a critical component of NCSSM’s mission

to educate academically talented students to become

state, national, and global leaders in science, technology,

engineering, and mathematics. I am thrilled that each year

we continue to increase the outstanding opportunities

NCSSM students have to participate in research.

Each edition of Broad Street Scientific features some

of the best research that our students conduct at NCSSM

under the guidance of our outstanding faculty and in

collaboration with researchers at major universities. For

thirty-five years, NCSSM has showcased student research

through our annual Research Symposium each spring and

at major research competitions such as the Regeneron

Science Talent Search and the International Science and

Engineering Fair. The publication of this journal provides

another opportunity to share with the broader community

the outstanding research being conducted by NCSSM

students.

I would like to thank all of the students and faculty

involved in producing Broad Street Scientific, particularly

faculty sponsor Dr. Jonathan Bennett, and senior editors

Olivia Fugikawa, Jason Li and Eleanor Xiao. Explore and

enjoy!

Dr. Todd Roberts, Chancellor

4 | 2019-2020 | Broad Street Scientific


WORDS from the EDITORS

Welcome to the Broad Street Scientific, NCSSM’s journal

of student research in science, technology, engineering,

and mathematics. In this ninth edition of the Broad Street

Scientific, we aim to showcase the breadth and depth of research

conducted by our students and encourage readers

to be scientifically literate individuals. In rapidly changing

times, it is important to keep up with new advancements

so that we can hold informed perspectives and make informed

decisions as a society. We hope you enjoy this

year’s edition!

This year’s theme is the composition of the human

body: a visual dive into the components that connect each

of us as one species. Human anatomy and physiology is the

subject of research for many scientific disciplines, ranging

from biology and chemistry to physics and engineering. A

deep understanding of human anatomy is key to promoting

health and wellness in our society, and the years of research

that have gone into studying the human body have

produced significant improvements in healthcare and the

quality of life. We chose the composition of the human

body as this year’s theme not just because of its importance

to research, but also because of the incredible scientific

images produced during the research process. Technology

has enabled us to acquire high resolution images of

many biological components, from tissues and individual

cells down to proteins and nucleic acids. The anatomical

images used in this year’s issue display the many levels of

biological organization. The biological levels of organization

order the different complexities of life from simplest

to most complex. In living organisms, these levels are:

atoms, molecules, organelles, cells, tissues, organs, organ

systems, organisms. The importance of every working

component cannot be understated; all individual parts of

a system must function properly to produce a working

whole. This aspect of anatomy and physiology reminds us

that we are all part of something bigger, just as how every

bit of research done, even by high schoolers, contributes to

a greater society of scientific thinkers.

We would like to thank the faculty, staff, and administration

of NCSSM for continuing to support and build

the scientific community that NCSSM represents. Now in

its 40th year, NCSSM continues to nurture a stimulating

academic environment that encourages motivated students

to apply their interests towards solving real-world

problems. To many, NCSSM serves as a symbol of passion

and determination in the next generation of young

people that will no doubt change the world. We would

like to give special thanks to Dr. Jonathan Bennett for his

invaluable support and guidance throughout the publication

process. We would also like to thank Chancellor Dr.

Todd Roberts, Dean of Science Dr. Amy Sheck, and Director

of Mentorship and Research Dr. Sarah Shoemaker.

Lastly, the Broad Street Scientific would like to acknowledge

Mr. Erik Troan, Founder and Chief Technology Officer at

Pendo, for speaking with us about his unique career path

and perspectives on STEM and entrepreneurship, and for

imparting important advice to young scientists and entrepreneurs

so that they, too, may tread fearlessly and find

their own path forward.

Olivia Fugikawa, Jason Li, and Eleanor Xiao

Editors-in-Chief

Broad Street Scientific | 2019-2020 | 5


BROAD STREET SCIENTIFIC STAFF

Editors-in-Chief

Olivia Fugikawa, 2020

Jason Li, 2020

Eleanor Xiao, 2020

Publication Editors

Henry Chen, 2021

Kaylene Eun, 2021

Daniel Jin, 2021

Esha Shakthy, 2021

Marina Takara, 2021

Ashley Wang, 2021

Biology Editors

Alisha Kamath, 2021

Ishaan Maitra, 2020

Akshra Paimagam, 2021

Chemistry Editors

Suraj Rao, 2020

Varun Varanasi, 2020

Andrew Zhen, 2021

Engineering Editors

Sahil Azad, 2021

Sriya Mantena, 2021

Mathematics and

Computer Science Editors

Alvin Chen, 2021

Aakash Kothapally, 2020

Physics Editors

Andrei Mistreanu, 2021

Will Staples, 2020

Faculty Advisor

Dr. Jonathan Bennett

6 | 2019-2020 | Broad Street Scientific


THE FORSAKEN VICTIMS OF CLIMATE CHANGE

Megan Mou

Megan Mou was selected as the winner of the 2019-2020 Broad Street Scientific Essay Contest. Her award included the

opportunity to interview Mr. Erik Troan '91, founder and CTO of Pendio.io. This interview can be found in the Featured

Article section of the journal.

Mother Earth is dying. Nowadays, it seems like there

is always a headline that bears news of yet another

environmental tragedy, and 2019 was a record year:

20% of the Amazon ravaged by forest fires, almost 900

deaths in the hottest England heat wave in history, and

1 billion animals killed by Australia’s wildfires added up

to the one of the most environmentally devastating years

in history. The worst part of it all? Humans are directly

linked to causing these catastrophes—yet we are still not

doing enough to prevent them. Research has shown that

humans are by far the greatest contributor of CO 2

, with

two thirds of human emissions due to the burning of fossil

fuels (Johnson, 2019). And when carbon dioxide and other

gases enter our atmosphere, they remain for centuries,

inflicting terrible damage on our planet.

Why do our activities have such a great effect on

Earth’s climate? To answer this, we must fully understand

the science behind global warming. The cause can be

summed up in one phrase: the enhanced greenhouse

effect. When solar radiation reaches Earth’s atmosphere,

some is reflected back into space, and the rest is absorbed

by the land and ocean, causing Earth to radiate heat

towards space. The greenhouse gases that naturally exist

in our atmosphere are responsible for absorbing some of

this heat, keeping Earth warm enough to sustain life ("The

Causes of Climate Change," 2019). The problem now is

that too much of these greenhouse gases (CO 2

, methane,

nitrous oxide, water vapor, and chlorofluorocarbons,

to name a few) is being released due to human activity,

driving up average annual temperatures. The recent U.S.

Fourth National Climate Assessment found that between

93% to 123% of observed 1951-2010 warming was due to

human activities (it is possible to be greater than 100%

because natural factors in isolation, such as volcanoes and

solar activity, would most likely have resulted in a slight

cooling over the last few decades, offsetting part of the

human-associated warming) (Hausfather, 2017). Most

people who are aware of climate change are also aware of

its general environmental impacts, such as melting glaciers

and rising sea levels, as well as disruption of habitats and

loss of biodiversity. However, what many are not aware of

is the disproportionately large impact that climate change

has on those living in extreme poverty.

There are several ways that climate change perpetuates

a cycle of poverty among the most vulnerable members

of our population. Poorer communities are not only more

susceptible to the detrimental effects of global warming,

they are also less able to cope with the aftermath of these

effects. If we continue to be bystanders, climate change

could force an additional 100 million people into extreme

poverty by 2030 (Giovetti, 2019). The most significant

threat that climate change poses to the poor is the

destruction of resources. 3 out of 4 people living in poverty

rely on agriculture and natural resources to survive—for

them, climate change is literally a matter of life and death.

So when increasingly unpredictable weather and natural

disasters such as sustained droughts or severe storms hit

farmers living in Haiti, Timor-Leste, or Zimbabwe, their

families face starvation and dehydration as their crops and

water sources are wiped out. Climate change threatens

the world’s food and water supply—research shows that

climate and natural disasters alone have triggered food

crises for 29 million people. The undermining of food

production and loss of resources cause extreme instability

in different communities, leading to a rise in conflicts over

arable land, freshwater, and livestock. In fact, thousands

have been killed in Nigeria, where diminishing land

and water resources have recently been exacerbated by

the effects of climate change, prompting long-standing

tension to spark into full-fledged violence (Schwartz,

2019). Clearly, existing social and economic problems are

only intensified by the consequences of climate change,

and millions of people are losing their lives as a result.

Those who have not been killed are being forced to flee.

Rising sea levels, prolonged droughts, and natural disasters

drive millions to move away from their homes in search of

better livelihoods. Almost all of these displacements occur

in developing countries, where residents have insufficient

resources to cope with change and destruction. This cycle

of poverty continues for refugees, many of whom do not

have access to education and must depend on humanitarian

aid. Furthermore, many first-world countries outsource

CO 2

in the form of factories to developing countries,

whose labor force is eager for new jobs (Plumer, 2017).

In short, climate change makes rich countries richer and

poorer countries poorer.

Broad Street Scientific | 2019-2020 | 7


Lastly, climate change multiplies the number of health

issues that exist in poorer regions. A warmer climate means

warmer freshwater sources, which in turn provide a more

habitable place for harmful bacteria and microbes to grow.

The World Health Organization (WHO) estimates that

3.575 million people die from water-related diseases per

year, and with increased temperatures drying out available

water sources, people driven desperate by thirst are forced

to choose between the risks of drinking contaminated

water or dying of thirst. Additionally, the increased smog

caused by warmer atmospheres, coupled with severe air

pollution, has made it impossible to breathe in places such

as Delhi, where the quality of air reached such high toxicity

that experts deemed it equivalent to smoking 50 cigarettes

a day (Paddison, 2020). In fact, the WHO claims that over

90% of the world population breathe in some form of

toxic air, leading to an abundance of diseases like stroke

and lung cancer (Fleming, 2018). Even within the U.S.,

poorer communities in both rural and urban areas bear the

greatest burden of climate change, as seen by lack of health

insurance, dependence on agriculture-based economies,

and no funds to recover from natural disasters. In urban

areas, which produce 80% of greenhouse gas emissions in

North America, the poor live in neighborhoods with the

greatest exposure to climate and extreme weather events

(Chappell, 2018). Poorer Americans, while to a much

lesser extent, face some of the same disadvantages as those

living in developing countries in terms of environmental

inequality. So what exactly is being done to save our planet

and its poorest inhabitants?

One thing is for sure: not enough. The overall global

response to climate change can be characterized as

extremely uneven. Persistent skepticism from certain

global leaders, many of whom are motivated by economic

interests, is slowing cooperative efforts to address the

issue of climate change. In particular, President Trump’s

decision to withdraw from the Paris climate agreement will

trigger both short-term and long-term damage—for one, it

will be less likely for the U.S., the second-highest ranking

country in production of greenhouse gases, to reduce

carbon emissions without international obligations, and

countries that were already hesitant about membership

are more likely to back off as well (“The Uneven Global

Response to Climate Change,” 2019). President Trump's

decision in times like these is brutal, as the urgency of

action is greater now than ever before, as indicated by the

World Meteorological Organization’s 2019 report. The

report showed a continued increase in greenhouse gases

to new records during the period of 2015-2019, with CO 2

growth rates nearly 20% higher than the previous five

years (“Global Climate in 2015-2019,” 2019). Surrounded

by a constant whirlwind of bad news, it can feel like all

hope is lost for planet Earth.

But hope is not lost. The global effort to address

climate change is moving forward as a whole, even

without the current support of the U.S. government.

Plenty of countries are setting goals to reduce emissions

and implement renewable energy. Morocco has invested

heavily in solar power, and India has implemented a

prohibition of new coal plants (“The Uneven Global

Response to Climate Change,” 2019). China, the world’s

number one CO 2

contributor at a whopping 29% of

global emissions, has made great strides in reducing air

pollution (“Each Country's Share of CO 2

Emissions,”

2019). The youth movement across multiple nations, led

by activist Greta Thunberg, is on the rise, and hundreds

of new green technologies are making their way towards

the market. The most promising of these innovations

include solar cells that incorporate the mineral perovskite,

which convert UV and visible light with a stunning 28%

efficiency (as compared to the average 15-20%), graphenebased

batteries that power electric vehicles, and carbon

capture and storage that traps CO 2

at its source and isolates

it underground (Purdue University, 2019).

While these global pledges and new technologies

hold great promise for future sustainability, it is up to us

to actively implement more environmentally conscious

decisions into our daily lives. Reduce, reuse, and recycle in

that order. Eat a more plant-based diet. Conserve energy

at every moment possible. Always be civically engaged.

We owe it to not only the animals, our children, and our

home; we owe it to those who contribute least to the

climate change devastation but feel its effects most deeply.

We must not let our privilege go wasted.

References

Chappell, C. (2018, November 27). Climate change in

the US will hurt poor people the most, according to a

bombshell federal report. Retrieved January 7, 2020, from

https://www.cnbc.com/2018/11/26/climate-changewill-hurt-poor-people-the-most-federal-report.html.

Each Country's Share of CO 2

Emissions. (2019, October

10). Retrieved January 8, 2020, from https://www.ucsusa.

org/resources/each-countrys-share-co2-emissions.

Fleming, S. (2018, October 29). More than 90% of

the world's children are breathing toxic air. Retrieved

January 9, 2020, from https://www.weforum.org/agenda/2018/10/more-than-90-of-the-world-s-children-arebreathing-toxic-air/.

Giovetti, O. (2019, September 25). How the effects of

climate change keep people in poverty. Retrieved January

7, 2020, from https://www.concernusa.org/story/effectsof-climate-change-cycle-of-poverty/.

Global Climate in 2015-2019: Climate change accelerates.

8 | 2019-2020 | Broad Street Scientific


(2019, September 24). Retrieved January 8, 2020, from

https://public.wmo.int/en/media/press-release/global-climate-2015-2019-climate-change-accelerates.

Hausfather, Z. (2017, December 13). Analysis: Why

scientists think 100% of global warming is due to humans.

Retrieved January 8, 2020, from https://www.carbonbrief.org/analysis-why-scientists-think-100-of-globalwarming-is-due-to-humans.

Johnson, S. (2019, November 22). What Are the Primary

Heat-Absorbing Gases in the Atmosphere? Retrieved January

7, 2020, from https://sciencing.com/primary-heatabsorbing-gases-atmosphere-8279976.html.

Paddison, L. (2020, January 3). 5 Environmental News

Stories To Watch In 2020. Retrieved January 9, 2020,

from https://www.huffpost.com/entry/environ-

ment-heat-wave-climate-change-elections_n_5def-

87d7e4b05d1e8a57cb90.

Plumer, B. (2017, April 18). How rich countries "outsource"

their CO 2

emissions to poorer ones. Retrieved

January 9, 2020, from https://www.vox.com/energy-and-environment/2017/4/18/15331040/emissions-outsourcing-carbon-leakage.

Purdue University. (2019, November 12). New material

points toward highly efficient solar cells. Retrieved

January 9, 2020, from https://www.sciencedaily.com/

releases/2019/11/191112164944.htm.

Schwartz, E. (2019, November 19). Quick facts: How

climate change affects people living in poverty. Retrieved

January 8, 2020, from https://www.mercycorps.org/articles/climate-change-affects-poverty.

The Causes of Climate Change. (2019, September 30).

Retrieved January 7, 2020, from https://climate.nasa.gov/

causes/.

The Uneven Global Response to Climate Change. (2019,

November 18). Retrieved January 7, 2020, from https://

www.worldpoliticsreview.com/insights/27929/the-uneven-global-response-to-climate-change.

Broad Street Scientific | 2019-2020 | 9


SMALL MOLECULE ACTIVATION OF

WNT/ β-CATENIN SIGNALING PATHWAY ON

NEURODEGENERATION RATES OF DOPAMINERGIC

NEURONS IN C. ELEGANS

Ariba Huda

Abstract

Parkinson’s Disease (PD) is a neurodegenerative disease characterized by loss of midbrain dopaminergic (mDA) neurons.

While there are several medical treatments available for PD, they often come with significant side effects and do not act

as definite cures. Past studies have indicated that Wnt/β-Catenin signaling is critical for the generation of dopamine

(DA) neurons during development and for further neurorepair. This study investigates the roles of small molecules,

Wnt Agonist 1 and Pyrvinium, in Wnt signaling and their effects on neurodegeneration. Wnt signaling was modeled

by Caenorhabditis Elegans (C. elegans), nematodes that display dopamine-dependent behavior in response to neurodegeneration.

24 hour exposure to Wnt Agonist 1 has been shown to significantly reduce neurodegeneration as observed

through locomotor behavior and chemotaxation. Currently, work is being done to measure BAR-1 and other Wnt related

ortholog gene expression within Wnt Agonist 1 exposed worms. Analysis of the functions behind the Wnt/β-Catenin

signaling pathway in the generation and neurorepair of mDA neurons will allow further understanding of the potential

for PD stem cell therapies.

1. Introduction

1.1 – Parkinson’s Disease

Parkinson’s Disease (PD) is a neurodegenerative disease

that results from the progressive cell death of dopaminergic

(DA) neurons. Currently, various medications

are prescribed to patients to control symptoms such as

cognitive decline and loss of motor function, but there is

no definitive cure. The most common therapy is the drug

levodopa (L-DOPA), which is used to stimulate dopamine

production in neurons associated with motor skill. However,

while L-DOPA is efficient at managing the extent

of symptoms, it also has various side effects on patients,

ranging from physiological to psychological. At present,

research is being conducted to identify the clinical applications

of stem cell therapy in Parkinson’s Disease as well as

the genetic factors behind PD [1]. These discoveries could

contribute to the development of a new therapy option

geared towards reducing the adverse effects of medications

on diagnosed patients.

Dopamine neurons are located in the human nigrostriatal

pathway, a brain circuit that connects neurons in the

substantia nigra pars compacta with the dorsal striatum.

Despite lack of a specific cause for neuronal loss, DA neuron

loss has been linked to genetic mutations and environmental

toxins [2]. Studies have shown that DA neurons

have a distinctive phenotype that could contribute to their

vulnerability. An example of this is the opening of L-type

calcium channels, which results in elevated mitochondrial

oxidant stress and susceptibility to toxins [2]. Moreover,

DA neurons are susceptible to degeneration because of

extensive branching and amounts of energy required to

transmit nerve signals along these branches [3].

1.2 – Wnt/β-Catenin Signaling Pathway

Due to the significance of the Wnt signaling pathway

for the healthy functioning of the adult brain, dysregulation

of these pathways in neurodegenerative disease has

become notable. Wnt/β-Catenin signaling is also critical

for the generation of DA neurons in embryonic stem cells

[4]. Since several of the biological functions disrupted in

PD are partially controlled by Wnt signaling pathways,

there is potential for therapy centered around targeting

these pathways [4].

In an activated state, Wnt proteins act as extracellular

signaling molecules that activate the Frizzled receptor.

Following the activation of Frizzled, the LRP receptor undergoes

phosphorylation, inducing the translocation of the

destruction complex, a complex of proteins that degrades

β-catenin, to the region of membrane near the two receptors

(Fig. 1). The activated dishevelled (Dsh) proteins cause

the inhibition of the destruction complex which prevents

β-catenin phosphorylation. Overall, an activated state of

the Wnt/β-Catenin signaling pathway causes an increase

in β-Catenin levels.

The transcription factor TCF mediates the genetic action

of Wnt signaling patterns, leading to the induction

of Wnt targeted genes. When β-Catenin levels increase,

they are translocated to the mitochondria, dislodging the

Groucho protein from TCF, and binding to TCF leading

to the transcription of Wnt targeted genes. The expressed

genes regulate cellular growth and proliferation. Without

10 | 2019-2020 | Broad Street Scientific BIOLOGY


Wnt stimulation, cytoplasmic β-Catenin levels are kept

low through continuous proteasome-mediated degradation.

Figure 1. Activated Wnt Signaling Pathway. Dsh is

activated when Wnt binds to the Frizzled receptor.

Consequently, the inhibition of the destruction complex

leads to increased expression of β-Catenin.

(GSK-3β), a protein found in the destruction complex

of the Wnt/β-Catenin signaling pathway [7]. The phosphorylation

of a protein by GSK-3β inhibits activity of its

downstream target. GSK-3β is active in a number of pathways,

including cellular proliferation, migration, glucose

regulation, and apoptosis.

1.3 – Wnt Agonist 1 and Pyrvinium

The most notable Wnt activators work by inhibiting

the GSK-3β enzyme found in the β-Catenin destruction

complex [7]. By inhibiting GSK-3β, Wnts disrupt the ability

of the complex to degrade β-Catenin, allowing β-Catenin

to accumulate in the nucleus and relay the Wnt signal

for transcription. Further, it has been shown that GSK-3β

dysregulation contributes to PD-like pathophysiology and

accumulation of alpha-synuclein [7]. Wnt Agonist 1 (Fig.

3) is a cell-permeable, small molecule agonist of the Wnt

signaling pathway. Wnt Agonist 1 induces accumulation of

β-Catenin by increasing TCF transcriptional activity and

altering embryonic development [8]. Studies have found

that the stimulation of Wnt/β-Catenin signaling pathway

with the Wnt agonist has been able to reduce organ injury

after hemorrhagic shock [9].

Recent studies have found that the canonical

Wnt/β-Catenin pathway is a key mechanism in controlling

DA neuron cell fate decision from neural stem cells

or progenitors in the ventral midbrain [5]. Wnt acts as a

morphogen which activates several signaling pathways,

specifically regulating the development and maintenance

of midbrain dopaminergic (mDA) neurons [6]. Wnt proteins

have also been shown to regulate the steps of the processes

DA neuron specification and differentiation (Figure

2). Previous research in mice found that the Wnt/β-Catenin

signaling pathway is required to rescue mDA neuron

progenitors and promote neurorepair [4].

Figure 3. Wnt Agonist 1 Structure, (C

19 H 19 ClN 4 O 3 ). A

crystalline solid. Formula weight: 350.4

Figure 2. The development of midbrain dopamine

neurons. mDA neurons arise from neural progenitors

in the ventral midline, and are divided into 3

separate phases: regional and neuronal specification

(phase I), early (phase II), and late differentiation

(phase III) [20].

New approaches in stem cell research have utilized developmental

molecules to program embryonic stem cells.

The key ingredient for this is glycogen synthase kinase

BIOLOGY

Pyrvinium (Fig. 4) has been found to be a small molecule

inhibitor of the Wnt signaling pathway through activation

of casein kinase 1a [10]. In most eukaryotic cell

types, the casein kinase 1 family of protein kinases are

enzymes that function as regulators of signal transduction

pathways. Studies have demonstrated that pyrvinium

binds to and activates CK1α, a part of the β-Catenin destruction

complex. CK1α members play a critical role in

Wnt/β-Catenin signaling, acting as both a Wnt activator

and Wnt inhibitor [10].

Clinical research in neurodegenerative diseases has not

been done with Wnt Agonist 1 and Pyrvinium. However,

Wnt Agonist 1 has been shown to be effective in vivo, de-

Broad Street Scientific | 2019-2020 | 11


creasing tissue damage in rats with ischemia-reperfusion

injury [21]. Pyrvinium salts have been shown to inhibit the

growth of cancer cells [10]. Additionally, Pyrvinium has

been shown to be an effective anthelmintic, a drug used to

treat infections of animals with parasitic worms [22].

the other hand, if worms are exposed to increasing concentrations

of Pyrvinium, they will display higher rates of

neurodegeneration due to CK1α activation and inhibited

Wnt/β-Catenin signaling. Furthermore, if worms are exposed

to respective treatments during development, their

neurodegeneration rates will increase or decrease more

drastically compared to adult worms who are not exposed

to any treatment.

Figure 4. Pyrvinium Structure, (C26H28N3 + ). Formula

Weight: 382.53

1.4 – C. Elegans Model

C. elegans is a popular model for neurodegenerative

research. The transparency of C. elegans makes it easy to

facilitate the study of specific neurons and genetic manipulation

[11]. C. elegans also present locomotor behavioral responses

to neurodegeneration. The identification of genes

that cause monogenic forms of PD allows for easy modeling

in C. elegans. Studying a C. elegans model of PD can provide

insight into the cellular and molecular pathways involved

in human disease. C. elegans are also able to be used

to identify disease markers and test potential treatments.

Outcome measures are used to detect disease modifiers

such as survival of dopamine neurons, dopamine dependent

behaviors, mitochondrial morphology and function,

and resistance to stress. Behavioral markers studied in C.

elegans to detect PD include basal slowing, ethanol preference,

area restricted searching, swimming induced paralysis,

and accumulation of α-synuclein [12]. Dopamine

neuron location sites are present within hermaphroditic

and male specific worms (Fig. 5). This project will further

investigate the role of C. elegans in neurodegenerative research

in place of previously used vertebrate models.

1.5 – Hypothesis

Due to the significant role Wnt/β-Catenin signaling

plays for DA neuron differentiation in development, activated

signaling within worms will decrease neurodegeneration

rates. Thus, if worms are exposed to increasing

concentrations of Wnt Agonist 1, they will display lower

rates of neurodegeneration due to activated Wnt/β-Catenin

signaling and increased β-Catenin expression. On

Figure 5. Top represents a hermaphrodite C. elegans

and bottom represents a male C. elegans with their respective

DA neuron sites. R&L stands for right and

left side. CEPD neurons are mechanosensory neurons.

ADE neurons are anterior deirid neurons. PDE

neurons are post embryonically born posterior deirid

neurons. All neurons have DOP-2 dopamine receptor.

R5a, R7a, and R9a neurons are male specific

sensory ray neurons [25].

2. Methods

This study consists of a preliminary experiment, two

main experiments, and a secondary experiment. The two

C. elegans strains used in this project were the OW13 and

N2 type worms. OW13 overexpressed alpha-synuclein

and displayed PD-like symptoms. N2 worms acted as the

control group, representing wildtypes with no genetic alterations.

The same treatments were administered to both

worms for each of the following experiments. Treatments

administered included three separate concentrations of

Wnt Agonist 1 and Pyrvinium, as well as dimethyl sulfoxide

(DMSO).

2.1 – C. elegans Maintenance

N2 and OW13 strains were purchased from the

Caenorhabditis Genetics Center (CGC) at the University

of Minnesota. The CGC is funded by the National Institute

of Health (P40 OD010440). Worms were placed on

nematode growth media (NGM) plates that were spotted

with approximately 30μl solution of LB Broth and

OP50, a strain of E. coli in the C. elegans diet. Worms were

placed onto new plates approximately every 48-72 hours.

C. elegans maintenance protocols were followed through

WormBook [13].

12 | 2019-2020 | Broad Street Scientific BIOLOGY


2.2 – Preparation

Wnt Agonist 1 Preparation: Stock Wnt Agonist 1 was

purchased from Selleckchem. 1mg Wnt Agonist 1 was dissolved

into 2.5851 mL, 0.5170 mL, and 0.2585 mL DMSO

for 1 mM, 5mM, and 10 mM concentrations, respectively.

Dilutions were stored in 4 o C.

Pyrvinium Preparation: Stock Pyrvinium was purchased

from Selleckchem. 1 mg Pyrvinium was dissolved

into 0.8685 mL, 0.1737 mL, and 0.0896 mL DMSO for 1

mM, 5 mM, and 10 mM concentrations, respectively. Dilutions

were stored in 4 o C.

2.3 – Preparation: Experimental Design

Activating Wnt Signaling: During each trial, 6 NGM

plates spread with different concentrations of treatment

were prepared for each strain of worm. L4 stages of each

strain (OW13 and N2) were exposed to three different

concentrations (1 mM, 5 mM, and 10 mM) of Wnt Agonist

1. Worms were kept on the plate for approximately

24 hours until they reached the adult stage. Then, approximately

half of the worms were transferred to perform the

thrashing assays and the remaining worms remained on

the plate until the next generation of worms were apparent.

Inhibiting Wnt Signaling: During each trial, 6 NGM

plates, spread with different concentrations of treatment,

were prepared for each strain of worm. L4 stages of each

strain (OW13 and N2) was exposed to three different

concentrations (1 mM, 5 mM, and 10 mM) of Pyrvinium.

Worms were kept on the plate for approximately 24 hours

until they reached the adult stage. Then, some worms were

transferred to perform the thrashing assays, and the rest of

the worms remained on the plate until the next generation

of worms were apparent.

Thrashing Assay: Agar Pads were prepared and approximately

30 worms were picked onto each pad. A drop of

M9 buffer was then added. Five minutes after transfer, the

number of body bends in 20s intervals was sequentially

filmed then counted for each of the worms on the assay

plate. This was done after 24 hour exposure to treatment

and when the next generation of worms (born to exposure)

had reached adult stage. Trials were run 3 times for

each treatment group and values were averaged.

2.4 – Statistical Measurements

Averages of locomotor behavior and chemotaxation

were calculated and recorded after experimentation. Data

were analyzed using unpaired Student t-tests with unequal

variance. This is due to differences in sample size per treatment

group. Error bars were calculated using standard error

of the mean (SEM).

3. Results

3.1 – Preliminary Data

To determine that OW13 and N2 strains displayed significant

differences in neurodegeneration, a preliminary

thrashing assay was conducted on OW13 and N2 worms

with exposure only to DMSO. These worms represented

a non-treatment group. A single thrash is a complete body

bend (Fig. 6).

Figure 6. Single thrash (body bend) within 20 second

interval. Movement follows from top to bottom.

We can conclude that the number of body bends per 20

second interval is an appropriate measure of neurodegeneration

and displays statistically significant differences between

N2 and OW13 strains of worms. OW13 worms

thrashed at a rate of 34.58 bends per 20 second interval and

N2 worms thrashed at a rate of 23.42 bends per 20 second

interval (Fig. 7). The OW13 worms with overexpressed

α-synuclein showed a significant increase in thrashing in

comparison to the wildtype N2 worms. Overexpression

of α-synuclein is the main marker of dopamine deficiency

and Parkinson’s Disease. This demonstrates that higher

rates of thrashing correlate to higher levels of dopamine

deficiency within the OW13 strain.

BIOLOGY

Broad Street Scientific | 2019-2020 | 13


Figure 7. No treatment thrashing score analysis: the

number of body bends per 20 second interval were

counted for 30 worms per each strain. Then the values

were averaged and graphed above. N2 worms

moved at an average of 23.42 body bends per interval,

with a standard deviation of 3.86, median of 23.5,

and standard error of the mean (SEM) of 0.79. OW13

worms moved at an average of 34.58 body bends per

interval, with a standard deviation of 4.44, median

of 35, and SEM of 0.811. Significance (p<0.05 = *) is represented

through bold connecting lines. This shows

that worms with Parkinson’s Disease (OW13) move

at a faster rate than wildtype (N2) worms. Error bars

± SEM.

4. Main Results

After determining the effectiveness of the analysis of

thrashing scores in measuring neurodegeneration, worms

were exposed to their own respective treatments (Tab.

1). These treatments were either Wnt Agonist 1 (1 mM,

5mM, or 10 mM concentration), Pyrvinium (1 mM, 5mM,

or 10mM concentration), or DMSO. The number of body

bends per 20 second interval were counted for 30 worms

per each strain and concentration. The values were averaged

and graphed below.

Table 1. Organization of worm groups and treatments

administered. (Key: 1 = 1mM treatment, 5 =

5mM treatment, 10 = 10mM treatment)

Wnt Agonist 1

Effect on

Cancer

OW13 A1, A5, A10 C1, C5,

C10

N2 B1, B5, B10 D1, D5,

D10

DMSO

(Control)

3 plates of

worms exposed

to equal concentration

3 plates of

worms exposed

to equal concentration

Exposure to Wnt Agonist 1 significantly decreased

thrashing rates in comparison to the control DMSO worms

(Fig. 8). However, this does not show the effect was concentration

dependent. Exposure to Pyrvinium at 5mM

and 10mM concentrations killed the worms. 1mM Pyrvinium

exposure also resulted in a significant decrease in

thrashing rates. In the N2 groups, there is no significant

difference between the treatment groups and the control

N2 group. Pyrvinium decreased thrashing in relation to

the 10mM Wnt Agonist 1 exposure, contradicting the

original hypothesis.

Next generation adult worms were collected on the

same treatment plates after approximately 5 days. This

study further aimed to observe any developmental significance

in activated Wnt/β-Catenin signaling within genetically

predisposed offspring.

Figure 8. Thrashing score analysis across various

treatments after 24 hour exposure: 10 different treatment

groups of 30 worms were measured: OW13

Wnt Agonist 1 (1mM) (μ = 24.615 , ̃x = 25, SEM= 0.475);

OW13 Wnt Agonist 1 (5mM) (μ = 20.625 , ̃x = 21, SEM=

0.739); OW13 Wnt Agonist 1 (10mM) (μ = 20.923 , ̃x = 21,

SEM= 0.630); OW13 Pyrvinium (1mM) (μ = 29.167, ̃x =

29, SEM= 0.632); N2 Wnt Agonist 1 (1mM) (μ = 25.375 , ̃x

= 25, SEM= 0.789); N2 Wnt Agonist 1 (5mM) (μ = 27.846

, ̃x = 28, SEM= 0.414); N2 Wnt Agonist 1 (10mM) (μ =

27.75 , ̃x = 28.5, SEM= 0.567); N2 Pyrvinium (1mM) (μ

= 23 , ̃x = 23.5, SEM= 0.833); OW13 DMSO (μ = 34.58 , ̃x

= 35, SEM= 0.811); N2 DMSO (μ = 23.42 , ̃x = 23.5, SEM=

0.705). Significance (p<0.05 = *, p<0.01 = **, p<0.001 =

***) is represented through dotted connecting lines.

Error bars ± SEM.

Similar to the adult worms (Fig. 8), exposure to Wnt

Agonist 1 significantly decreased thrashing rates in comparison

to the control DMSO next generation worms (Fig.

9). However, there is no significance supporting a concentration

dependent effect. Exposure to Pyrvinium at 5mM

and 10mM concentrations killed the worms and 1mM

Pyrvinium exposure also resulted in significant decrease in

thrashing rates. In the N2 groups, there is no strong significance

between the treatment groups and the control N2

group. Pyrvinium decreased thrashing in relation to the

14 | 2019-2020 | Broad Street Scientific BIOLOGY


10mM Wnt Agonist 1 exposure. The Pyrvinium results

might be related to the dual function Pyrvinium has in

both activating and inhibiting Wnt signaling.

Wnt Agonist 1, and 1mM Pyrvinium were significant (Fig.

10). Next generation OW13 worms thrashed less when exposed

to 1 mM Wnt Agonist 1 and more when exposed

to 5 mM Wnt Agonist 1, but less for both in N2 strains.

These inconsistent results lead us to conclude that there

is no overall significant difference or correlation between

adult worms exposed to treatments and their offspring.

5. Discussion

Figure 9. Thrashing score analysis across various

treatments of next generation worms, born into

treatment exposure: 10 different treatment groups

of 30 worms were measured: OW13 Wnt Agonist 1

(1mM) (μ = 20.929 , ̃x = 21, SEM= 0.793); OW13 Wnt Agonist

1 (5mM) (μ = 20.438 , ̃x = 20, SEM= 0.653); OW13

Wnt Agonist 1 (10mM) (μ = 24.333 , ̃x = 21, SEM= 0.328);

OW13 Pyrvinium (1mM) (μ = 22.769 , ̃x = 22, SEM= 0.553);

N2 Wnt Agonist 1 (1mM) (μ = 21.75 , ̃x = 25, SEM= 0.405);

N2 Wnt Agonist 1 (5mM) (μ = 21.25 , ̃x = 21, SEM= 0.603);

N2 Wnt Agonist 1 (10mM) (μ = 25.2 , ̃x = 25, SEM= 0.618);

N2 Pyrvinium (1mM) (μ = 26.5 , ̃x = 236.5, SEM= 0.496);

OW13 DMSO (μ = 34.58 , ̃x = 35, SEM= 0.811); N2 DMSO

(μ = 23.42 , ̃x = 23.5, SEM= 0.705). Significance (p<0.05

= *, p<0.01 = **, p<0.001 = ***) is represented through

dotted connecting lines. Error bars ± SEM.

Using the above data, adult thrashing scores and next

generation thrashing scores were compared per strain.

Figure 10: Adult versus next generation worm

thrashing score analysis: Data collected in both Figure

10 and 11 were replotted adjacent to each other in

order for comparison between adult and next generation

worms of each strain. Significance (p<0.05 = *,

p>0.05 = ns) is represented through dotted connecting

lines. Error bars ± SEM.

The thrashing differences between adult and next generation

worms exposed to 1mM Wnt Agonist 1, 5mM

BIOLOGY

5.1 – Wnt/β-Catenin Signaling Activation on Neurodegenerative

Rates

Previous research has shown that Wnt/β-Catenin signaling

is critical for the generation of dopamine neurons in

embryonic stem cells [4]. The generation of DA neurons

increases dopamine levels in the brain, thus decreasing

neurodegeneration.

The number of body bends per 20 second interval is an

appropriate measure of neurodegeneration, and displays

statistically significant differences between N2 and OW13

strains of worms (Fig. 7). This further shows that higher

values of body bends per 20s interval correlate to increased

neurodegeneration.

Wnt Agonist 1 effectively decreases neurodegeneration

rates in adult and next generation worms in comparison

to worms exposed to DMSO (Fig. 8 & 9). This decrease

in neurodegeneration, however, is not significant in a

concentration dependent manner. Pyrvinium in a 1mM

concentration has also been shown to decrease neurodegeneration,

which differs from the original hypothesis.

This could be due to the dual function of CK1α members

as Wnt activators or Wnt inhibitors. Pyrvinium activates

CK1α in the β-catenin destruction complex. Further, there

is no consistently significant difference in neurodegeneration

rates between adult and next generation worms (Fig.

10). This also differs from the original hypothesis which

suggests that Wnt activation might be a more effective

treatment during development.

5.2 – Limitations

Wnt/β-catenin signaling has been suggested to be

different in C. elegans than in vertebrates. In metazoans

(cnidarians, nematodes, insects and vertebrates), Wnts

are secreted glycoproteins that function as extracellular

signals [23]. Evolutionary conservation suggests that cell

signaling functioning in response to Wnts were part of a

“developmental toolkit” from at least 500 million years ago

from the common ancestor to modern metazoans [23]. Although

C. elegans uses Wnt/β-catenin signaling similar to

other metazoans, it also has a second Wnt/β-catenin signaling

pathway that uses extra β-catenins for up-regulation

of target genes, distinct from other species [23]. This

could lead to varied results in activated Wnt signaling in

comparison to other organisms.

In order to conduct thrashing assay, adult worms were

Broad Street Scientific | 2019-2020 | 15


picked from the treatment plates to agar pads spotted with

M9 buffer. Often worms were picked incorrectly and died

very quickly. Thrashing scores of these worms were discarded

from analysis.

6. Current Work

Expression patterns of Wnt related genes during larval

development have been extensively studied using transgenic

reporter gene based assays. Wnts have been established

to act as morphogens, providing cells in developing tissue

with positional information in long-range concentration

gradients [14]. Sawa and Korswagen [14] looked at Wnt

related genes, BAR-1, POP-1, GSK-3, PRY-1, MOM-2,

MOM-5, KIN-19, which are orthologs of significant genes

that partake in Wnt/β-Catenin signaling. BAR-1 (β-catenin/armadillo-protein

1) functions as a transcriptional activator,

and along with POP-1 (ortholog of Tcf), regulates

cell fate decision during larval development [15]. GSK-3

(Glycogen synthase kinase-3) is the ortholog of human

GSK3β, a key enzyme in Wnt signaling and phosphorylation

of β-catenin [16]. PRY-1 is the ortholog of Axin-1

and is a part of the destruction complex in negatively regulating

BAR-1/β-catenin signaling [17]. MOM-2 codes a

Wnt ligand for members of the Frizzled family as well as

regulates cell fate determination [18]. MOM-5 (ortholog

of Frizzled receptor) couples to the β-catenin signaling

pathway, leading to the activation of disheveled proteins

[14]. KIN-19, ortholog of CK1 (Casein Kinase 1), has been

shown to transduce Wnt signals [19].

Future work in this study will be done to extensively

look at Wnt related gene expression within worms that

show decreased neurodegeneration. This will be done

through cDNA synthesis and real time polymerase chain

reaction. We expect to see BAR-1 expression, the ortholog

of β-catenin, and other Wnt related genes to have increased

expression within worms exposed to Wnt Agonist

1 in all concentrations. However, we expect to see less expression

of GSK-3 as decreased β-catenin is expected to be

degraded through phosphorylation.

7. Conclusions

Currently, there are no disease modifying treatments

for Parkinson’s Disease. Current PD treatments involve

the use of dopaminergic drugs to restore dopamine concentration

and motor function. These treatments do not

alter the course of PD, but they do provide improvement

in motor symptoms of patients [1]. Numbers of cell-based

treatments have responded to the need for targeted delivery

of physiologically released dopamine. One option that

recent studies have considered is the introduction of stem

cells into the striatum [1]. Lineage tracing based on Wnt

target genes has provided evidence for Wnts as significant

stem cell signals that have been detected in various organs

[16]. Wnt proteins or Wnt agonists have been used to

maintain stem cells in culture, allowing stem cells to expand

in a self-renewing state [24].

Though C. elegans do not have the same Wnt/β-catenin

signaling system as vertebrates, they are a valuable model

to test whether or not Wnt targeted therapies are effective

treatments to increase dopamine production in neurons

and decrease PD symptoms. The use of Wnt activators on

model organisms have not been well studied, especially in

the context of neurodegenerative disease.

As more studies and trials are completed on the effects

of activated Wnt/β-catenin signaling, especially through

the exposure to various agonists, we can see how organisms

respond physiologically, genetically, and behaviorally

to such changes. Further experimentation should also

consider potential side effects of such treatments as well

as the toxicity of molecules used for activation. Analysis

should further study if Wnt signaling is able to rescue neurodegeneration

by inducing DA neuron development or

through neurorepair. The most effective clinical treatment

of Parkinson’s disease can be achieved by expanding the

field and examining potential therapies.

8. Acknowledgements

I would like to thank my mentor, Dr. Kim Monahan,

and my Research in Biology class for guiding and supporting

me through the research process. I would also like to

thank Angelina Katsanis and Emile Charles for being my

lab assistants over the summer. Further thanks to Dr. Amy

Sheck, the Glaxo Endowment, and the North Carolina

School of Science and Mathematics for allowing me the

opportunity to experience research.

9. References

[1] Stoker, T. B., & Greenland, J. C. (2018). Parkinson’s

Disease: Pathogenesis and Clinical Aspects. Codon Publications.

[2] Surmeier, D. J., Guzmán, J. N., Sánchez-Padilla, J., &

Goldberg, J. A. (2010). What causes the death of dopaminergic

neurons in Parkinson’s disease?. In Progress in

brain research (Vol. 183, pp. 59-77). Elsevier.

[3] Mamelak, M. (2018). Parkinson’s disease, the dopaminergic

neuron and gammahydroxybutyrate. Neurology

and therapy, 7(1), 5-11.

[4] L'Episcopo, F., Tirolo, C., et al. (2014). Wnt/β‐catenin

signaling is required to rescue midbrain dopaminergic progenitors

and promote neurorepair in ageing mouse model

of Parkinson's disease. Stem Cells, 32(8), 2147-2163.

[5] Chen, L. W. (2013). Roles of Wnt/β-catenin signaling

16 | 2019-2020 | Broad Street Scientific BIOLOGY


in controlling the dopaminergic neuronal cell commitment

of midbrain and therapeutic application for Parkinson’s

disease. Trends In Cell Signaling Pathways In Neuronal

Fate Decision, 141.

[6] Arenas, E. (2014). Wnt signaling in midbrain dopaminergic

neuron development and regenerative medicine for

Parkinson's disease. Journal of molecular cell biology, 6(1),

42-53.

[7] Credle, J. J., George, J. L., et al. (2015). GSK-3 β dysregulation

contributes to Parkinson’s-like pathophysiology

with associated region-specific phosphorylation and

accumulation of tau and α-synuclein. Cell Death & Differentiation,

22(5), 838-851.

[8] Wnt Agonist I (CAS 853220-52-7). (n.d.). Retrieved

from https://www.caymanchem.com/product/19903/

wnt-agonist-i.

[9] Kuncewitch, M., Yang, W. L., et al. (2015). Stimulation

of Wnt/beta-catenin signaling pathway with Wnt agonist

reduces organ injury after hemorrhagic shock. The journal

of trauma and acute care surgery, 78(4), 793.

[10] Thorne, C. A., Hanson, A. J., et al. (2010). Small-molecule

inhibition of Wnt signaling through activation of casein

kinase 1α. Nature chemical biology, 6(11), 829.

[11] Wolozin, B., Gabel, C., Ferree, A., Guillily, M., &

Ebata, A. (2011). Watching worms whither: modeling

neurodegeneration in C. elegans. In Progress in molecular

biology and translational science (Vol. 100, pp. 499-514).

Academic Press.

[12] Vartiainen, S. U. V. I., & Wong, G. A. R. R. Y. (2005).

C. Elegans models of parkinson disease.

[13] Wang, D., Liou, W., Zhao, Y., & Ward, J. (2019, September

3). Retrieved from http://www.wormbook.org/

chapters/www_strainmaintain/strainmaintain.html.

The Axin-like protein PRY-1 is a negative regulator of a

canonical Wnt pathway in C. elegans. Genes & Development,

16(10), 1291-1302.

[18] mom-2 (gene) - WormBase : Nematode Information

Resource. (n.d.). Retrieved from https://wormbase.org/

species/c_elegans/gene/WBGene00003395#0-9g-3.

[19] Peters, J. M., McKay, R. M., McKay, J. P., & Graff,

J. M. (1999). Casein kinase I transduces Wnt signals.

Nature, 401(6751), 345-350.

[20] Ferri, A. L., Lin, W., Mavromatakis, Y. E., Wang, J.

C., Sasaki, H., Whitsett, J. A., & Ang, S. L. (2007). Foxa1

and Foxa2 regulate multiple phases of midbrain dopaminergic

neuron development in a dosage-dependent

manner. Development, 134(15), 2761-2769.

[21] Kuncewitch, M., Yang, W., Nicastro, J., Coppa, G.

F., & Wang, P. (2013). Wnt Agonist Decreases Tissue

Damage and Improves Renal Function After Ischemia-Reperfusion.

Journal of Surgical Research, 179(2),

323.

[22] Holden-Dye, L., & Walker, R. (2014). Anthelmintic

drugs and nematocides: studies in Caenorhabditis elegans.

WormBook: the online review of C. elegans biology,

1-29.

[23] Jackson, B. M., & Eisenmann, D. M. (2012). β-catenin-dependent

Wnt signaling in C. elegans: teaching an

old dog a new trick. Cold Spring Harbor perspectives in

biology, 4(8), a007948.

[24] Zeng, Y. A., & Nusse, R. (2010). Wnt proteins are

self-renewal factors for mammary stem cells and promote

their long-term expansion in culture. Cell stem cell, 6(6),

568-577.

[25] (n.d.). Retrieved from http://home.sandiego.edu/~cloer/loerlab/dacells.html.

[14] Sawa, H. and Korswagen, H. C. (2013). Wnt signaling

in C. elegans. Retrieved from http://www.wormbook.

org/chapters/www_wntsignaling.2/wntsignal.html

[15] BAR-1 (protein) - WormBase : Nematode Information

Resource. (n.d.). Retrieved from https://wormbase.

org/species/c_elegans/protein/CE08973#062--10.

[16] Main. (n.d.). Retrieved from https://web.stanford.

edu/group/nusselab/cgi-bin/wnt/.

[17] Korswagen, H. C., Coudreuse, D. Y., Betist, M. C.,

van de Water, S., Zivkovic, D., & Clevers, H. C. (2002).

BIOLOGY

Broad Street Scientific | 2019-2020 | 17


QUORUM QUENCHING: SYNERGISTIC EFFECTS

OF PLANT-DERIVED COMPOUNDS ON BIOFILM

FORMATION IN VIBRIO HARVEYI

Preeti Nagalamadaka

Abstract

Sixteen million people die annually due to diseases caused by antibiotic resistant bacteria, sixty-five percent of which form

biofilms. Biofilms offer one thousand times more resistance to antibiotics with their exopolysaccharide matrix. Many bacteria,

including Vibrio harveyi, a model for Vibrio cholera, opt for this structure to evade antibiotics. Because biofilm matrix

production is regulated by quorum sensing, efforts are underway to find quorum quenchers. This project focused on testing

combinations of plant-derived quorum quenchers that function by different mechanisms to find if they were more effective

than individual compounds at inhibiting biofilm formation in Vibrio harveyi. In previous literature, neohesperedin

and naringenin were found to inhibit HAI-1 and AI-2 signaling. Cinnamaldehyde also disrupted the DNA-binding ability

of the regulator LuxR. V. harveyi biofilms were grown in the presence of quorum quenching compounds with dimethyl

sulfoxide as a control, stained with Crystal Violet and quantified by OD. Naringenin alone was found to decrease biofilm

formation, whereas cinnamaldehyde and neohesperedin alone showed no detectable effect. Combinations of naringenin

and cinnamaldehyde showed a synergistic effect on inhibiting biofilm formation. Through studying V. harveyi, optimized

quorum quenching could be utilized to counter V. cholera and other biofilm-spread diseases.

1. Introduction

To conserve energy, bacteria coordinate metabolically

expensive activities through quorum sensing. Small

amounts of bacteria bioluminescing are metabolically

wasteful because it will not produce significant light, but

a large group of coordinated bacteria bioluminescing simultaneously

has an ecologically stronger effect, conserving

energy and benefitting all bacteria. Bacteria use chemical

signals to communicate with each other and induce

changes in the bacterial population. When there is a high

cell density of bacteria, molecules called autoinducers are

produced. Upon reaching a threshold concentration, the

whole bacterial population is signaled to alter its gene expression

in unison – a process called quorum sensing [1].

Autoinducers collectively control the activity of metabolically

expensive bacterial functions such as biofilm formation,

pathogenesis, bioluminescence, conjugation and

secretion of virulence factors [1]. In many species such

as Pseudomonas aeruginosa, Helicobacter pylori, Vibrio fischeri,

Vibrio cholerae and Vibrio harveyi, quorum sensing is a

means for bacterial survival and host pathogenesis [1][2].

Some bacteria use a single type of autoinducer in quorum

sensing, usually acyl homoserine lactones (AHLs) or

autoinducing peptides (AIPs) [3], while others use many

types of autoinducers. The single autoinducer LuxIR system

is common in many pathogenic gram-negative bacteria,

but the systems present in V. harveyi, P. aeruginosa and

V. cholerae differ because they have multiple components

and autoinducers. V. harveyi respond to three different

autoinducers (Fig. 1): V. harveyi autoinducer-1 (HAI-1),

Cholera autoinducer-1(CAI-1), and Autoinducer-2 (AI-2)

[4]. HAI-1 is a homoserine lactone (HSL) N-(3-hydroxybutanoyl)-HSL

which is a type of AHL. CAI-1 is a 3-hydroxytridecan-4-one,

and AI-2 is a furanosylborate diester

[4]. AI-2 is found in quorum sensing pathways among different

species and is thought to contribute to interspecies

communication. HAI-1, CAI-1, and AI-2 are recognized

by the sensor kinases LuxN, CqsS, and LuxQ/P respectively

[4]. The low concentration signal is received at these

receptors and is transduced by the phosphorylation phosphotransferase

LuxU which then phosphorylates LuxO

[5]. This activates the transcription of small regulatory

RNAs (sRNAs) that prevent the translation of LuxR [5].

The LuxR protein then goes on to regulate the expression

of over hundreds of genes involved in biofilm formation,

virulence factor secretion or bioluminescence. At

higher concentrations of autoinducers LuxN, LuxQ, and

LuxP switch to phosphatases, dephosphorylating LuxO.

Since dephosphorylated LuxO is inactive, no sRNAs will

be formed and thus the LuxR mRNA will be stable and

translated [6].

The quorum sensing pathway was first observed in

the bioluminescent species V. harveyi and is responsible

for regulating its bioluminescence, colony morphology,

biofilm formation and virulence factor production [7].

Biofilms are communities of bacteria stuck to a surface

encapsulated in an exopolysaccharide matrix. The gene

responsible for the production of this matrix is regulated

by the LuxR protein. Due to their exopolysaccharide

coats, bacteria from biofilms can evade the host immune

system and survive longer in harsh environments, leading

to critical economic problems and nosocomial infections.

Because quorum sensing is integral to the survival of bio-

18 | 2019-2020 | Broad Street Scientific BIOLOGY


films, research is underway to inhibit it, termed quorum

quenching [1][3][8]. To detect the efficacy of quorum

quenching, biofilm mass and certain virulence factors can

be quantified.

Many quorum sensing pathogens like Vibrio cholerae,

Providencia stuartii, Heliobacter pylori and Candida albicans

cause harmful infections in humans [2][11][12]. Quorum

sensing pathways in these species are analogous to pathways

in V. harveyi. Quenching these pathways can be impactful

because it offers an alternative approach to target

bacterial infections via quorum sensing [2][11]. Quorum

quenching works by competitively inhibiting the receptor

sites of the autoinducers, degrading autoinducers, or stopping

production of autoinducers completely [3][12]. The

most effective quorum quenchers are those that inhibit

biofilm formation and virulence factor secretion without

slowing the growth of bacteria, because inhibiting bacterial

growth can lead to more selective pressure and resistance

[8].

Many quorum quenching agents have been found in

marine organisms and herbal plants [2][11][12]. Some

specific plant compounds found to inhibit biofilms are

cinnamaldehyde (derived from cinnamon bark extract),

neohesperidin and naringenin (both derived from citrus

extracts) [9][10]. Cinnamaldehyde works by decreasing

the DNA binding ability of the LuxR transcript. Neohesperidin

inhibits the efficacy of the HAI-1 autoinducer in V.

harveyi. Naringenin inhibits quorum sensing via the AI-2

and HAI-1 autoinducer pathways. Because all three quorum

quenchers have different mechanisms of actions in

quorum quenching, combinations of these molecules were

tested on V. harveyi for possible synergistic effects in the

reduction of biofilm formation.

2. Materials and Methods

Figure 1. Shows the HAI-1, CAI-1 and AI-2 autoinducers

received by kinases/phosphatases LuxN, CqsS

and LuxP/Q respectively. The signal transduction

pathway regulating the expression of LuxR is also

shown via the LuxR transcript. LuxR functions in

the regulation of quorum sensing controlled behaviors

[4].

Quorum quenching can have a wide range of effects

in medicine. It can be responsible for inhibiting biofilm

formation and decreasing the pathogenicity of bacteria.

Because quorum sensing controls biofilm formation and

virulence factor secretion, it has the potential to decrease

pathogenicity. For example, the quorum sensing regulated

type three secretion system found in many gram-negative

bacteria infects eukaryotic cells via the secretion of specific

proteins. The quorum quencher naringenin has been

shown to decrease the virulence of the type three secretion

system [9]. Furthermore, the same quorum sensing pathways

control the secretion of exopolysaccharides needed

for maintaining biofilm structure [10].

BIOLOGY

2.1 – Compounds

Cinnamaldehyde, naringenin and neohesperidin were

purchased from Sigma-Aldrich. All compounds were dissolved

in dimethyl sulfoxide (DMSO) at a concentration of

10 mg/mL and stored at -20°C.

2.2 – Media and Bacterial Growth

Vibrio harveyi strain BB120 (wild-type) was purchased

from ATCC. The Luria Marine (LM) medium was used

to grow V. harveyi. Overnignt cultures were incubated at

30°C without shaking.

2.3 – Individual Compound Biofilm Formation Assay

This experiment determined the effect of each of the

quorum quenchers individually. Overnight culture of V.

harveyi BB120 was diluted in a 1:100 ratio in LM media.

One mL of the diluted culture was placed into each well

in a 24-well plate. Wells received concentrations of compounds

previously found to successfully inhibit quorum

sensing in V. harveyi [9][10]. Each well received 6.25 µg/

mL of naringenin, 12.5 µg/mL of neohesperidin, 13.22 µg/

mL of cinnamaldehyde, or 1.32 µL of DMSO as a control.

The plates were incubated at 26°C with no shaking for 24

Broad Street Scientific | 2019-2020 | 19


hours to stress bacteria into forming biofilms [9]. The biofilm

mass was quantified by Crystal Violet staining. The

plates were first washed with deionized water three times.

They were then stained with 2 mL of 0.1% Crystal Violet

solution for 20 minutes. The dye not associated with the

biofilm was washed out with deionized water. All the dye

associated with the biofilm was dissolved in 1 mL of 33%

acetic acid. The absorbances of these acetic acid and dye

samples were taken at 570 nm by a spectrophotometer.

The optical density (OD) was used as a means to quantify

the biofilm. This experiment was carried out 6 times with

6 replicated wells per plate.

2.4 – Multiple Compounds Biofilm Formation Assay

This experiment determined the effect of combinations

of quorum quenching compounds. This was the same assay

as Individual Compound Biofilm Formation Assay except

each well received 1.32 µL of DMSO as a control, 6.25

µg/mL of naringenin,12.5 µg/mL of neohesperidin, 13.22

µg/mL of cinnamaldehyde, or a combination of the compounds.

Combinations tested include cinnamaldehyde/

naringenin and neohesperidin/naringenin. Half concentrations

of naringenin (3.125 µg/mL) and cinnamaldehyde

(6.61 µg/mL) were tested in later combinations to determine

whether a synergistic effect was present. Plates were

stained with Crystal Violet and the OD associated with

biofilm was measured. Each combination experiment was

replicated 3 times with 4 replicated wells per plate.

2.5 – Cellular Growth Assay

This experiment determined whether combinations of

compounds altered the growth rates of the bacteria. Overnight

culture of V. harveyi BB120 was diluted in a 1:100 ratio

in Luria Marine (LM) media. Each tube received either

3.97 µL of DMSO as a control, 6.25 µg/mL of naringenin

and 12.5 µg/mL of neohesperidin, 6.25 µg/mL of naringenin

and 13.22 µg/mL of cinnamaldehyde, or 6.25 µg/mL

of naringenin, 12.5 µg/mL of neohesperidin and 13.22 µg/

mL of cinnamaldehyde. The cultures were grown for 24

hours at 30°C with shaking. Optical densities of the broth

were taken roughly every 2 hours for the first 8 hours.

Additionally, samples were taken of the broth at 6 hours

and 24 hours. The samples were serially diluted and plated

on LM agar plates and colony forming units were counted

after 24 hours of growth. This experiment was replicated

twice.

3. Data Analysis

To analyze the results of my experiments, the statistical

software JMP 10 was used to run ANOVA tests and

first determine the presence of a treatment effect. If the

ANOVA test showed an effect, a Tukey’s Honest Significance

Difference (HSD) test was conducted to discern the

significant differences among means. Some data were ad-

ditionally graphed as percent biofilm inhibition, further

emphasizing the presence of a synergistic effect. Biofilm

inhibition percentages were calculated as (control OD -

treatment OD)/(control OD)*100.

4. Results

4.1 – Individual Compound Biofilm Formation Assay

To measure the effectiveness of individual plant-derived

compounds on quorum sensing inhibition, the individual

compound biofilm formation assay was conducted

on V. harveyi biofilms. As determined by the Tukey’s honest

significance test, naringenin at a concentration of 6.25

µg/mL significantly decreased biofilm mass. Cinnamaldehyde

and neohesperidin showed no detectable biofilm inhibition

(Fig 2).

Figure 2. Individual Compound Biofilm Formation

Assay. Shows the results of the individual plant-derived

compounds on biofilm formation in V. harveyi

to see which compounds were most effective. Error

bars show mean +/- 1 SEM. An ANOVA test was conducted

and showed p < 0.0037* (Table 1). A Tukey’s

Significance test was then conducted. Different

letters correspond to significant differences as discerned

by Tukey’s HSD test.

Table 1. ANOVA table from the Individual Compound

Biofilm Formation Assay (Fig. 2).

Source DF Sum of

Squares

Compound

Mean

Square

F Ratio

Prob > F

3 0.34467 0.11489 6.2299 0.0037*

Error 20 0.36883 0.018442

Corr.

Total

23 0.7135

4.2 – Multiple Compounds Biofilm Formation Assay

Because naringenin was consistently effective, combinations

with naringenin were tested. The cinnamaldehyde

and naringenin combination (Fig. 3a) showed a significant

20 | 2019-2020 | Broad Street Scientific BIOLOGY


difference between the control and the individual compounds,

but no significant difference between the combination

and the individual compounds, which is indicative

of naringenin overpowering the combination.

Table 2. ANOVA test conducted on the combination

of naringenin and cinnamaldehyde (Fig. 3a).

Source DF Sum of

Squares

Mean

Square

F

Ratio

Prob > F

a)

3 0.34688 0.115628 12.0778 0.0002*

Error 17 0.16275 0.009574

Corr.

Total

20 0.50963

Table 3. ANOVA test conducted on the combination

of neohesperidin and naringenin (Fig. 3b).

Source DF Sum of

Squares

Mean

Square

F

Ratio

Prob > F

b)

Compound

Compound

3 0.41184 0.115628 17.9586 0.0001*

Error 20 0.15289 0.009574

Corr.

Total

23 0.56473

Figure 3. Combinations with naringenin. a) Shows

V. harveyi biofilm formation when treated with the

combination of cinnamaldehyde and naringenin.

Naringenin concentrations were 6.25 µg/mL and

cinnamaldehyde concentrations were 13.22 µg/mL.

The ANOVA test showed p < 0.0002* (Table 2), indicating

that there is indeed some difference between

the compounds. The Tukey’s HSD test was then conducted

to find specific differences. b) Shows V. harveyi

biofilm formation when treated with the combination

of naringenin and neohesperidin. The ANOVA

test showed p < 0.0001* (Table 3), indicating that there

is indeed some difference between the compounds.

Tukey’s HSD test was then conducted to find specific

differences. Different letters on the graphs above

indicate differences in significance according to

Tukey’s HSD test.

The neohesperidin and naringenin combination (Fig.

3b) showed naringenin significantly different from the

control and neohesperidin indistinguishable from the

control. Furthermore, the combination of naringenin and

neohesperidin was indistinguishable from naringenin,

supporting the theory that neohesperidin showed no detectable

effect. In these combinations the raw OD values of

the combinations were low, so in subsequent experiments

biofilm formation was increased through less volume of

LM media in all the wells to better detect an inhibitory

effect of the combinations. Since naringenin seemed

to overpower combinations, half the concentration of

naringenin (3.125 µg/mL) was tested with cinnamaldehyde

(Fig. 4). The combination with full concentrations

of naringenin and cinnamaldehyde showed a significant

difference in biofilm formation compared to the other

individual compounds, suggesting an interaction between

the compounds. Because this combination had a greater

percent biofilm inhibition than the individual compounds’

percents of inhibition combined (Fig. 5), it was concluded

that a synergistic effect was present. However, the

combination with half the concentration of naringenin

and full concentration of cinnamaldehyde was indistinguishable

from the full concentration of naringenin alone,

showing the necessity of naringenin for the synergy (Fig.

5). A similar experiment was conducted to determine the

effects of cinnamaldehyde concentration on this synergy

by testing combinations of naringenin with half the concentration

of cinnamaldehyde (6.61 µg/mL). A synergistic

effect was present in the combinations of naringenin

with both full and half concentrations of cinnamaldehyde,

indicating that the concentration of cinnamaldehyde is

not as integral as that of naringenin for the synergy (Fig.

6 and 7).

BIOLOGY

Broad Street Scientific | 2019-2020 | 21


Figure 4. Combinations with Varied Concentrations

of Naringenin. Effects on biofilm formation of the

combination of naringenin at full (6.25 µg/mL) and

half (3.125 µg/mL) concentrations with cinnamaldehyde

(13.22 µg/mL). Error bars show 1 SEM. ANOVA

test p < 0.0001* (Table 4). Letters correspond to differences

discerned by Tukey’s HSD Test.

Table 4. ANOVA test conducted on full concentrations

of cinnamaldehyde with varied concentrations

of naringenin (Fig. 4).

Source DF Sum of

Squares

Compound

Mean

Square

F

Ratio

Prob > F

5 0.94506 0.189011 29.7551 <0.0001*

Error 18 0.11434 0.006352

Corr.

Total

23 1.0594

Figure 6. Combinations with Varied Concentrations

of Cinnamaldehyde. Effects on biofilm formation of

the combination of cinnamaldehyde at full (13.22 µg/

mL) and half (6.61 µg/mL) concentrations with naringenin

(6.25 µg/mL). Error bars show 1 SEM. ANOVA

test p < 0.0001* (Table 5). Letters correspond to differences

discerned by Tukey’s HSD Test.

Table 5. ANOVA test conducted on full concentrations

of naringenin tested with various concentrations

of cinnamaldehyde (Fig. 6).

Source DF Sum of

Squares

Compound

Mean

Square

F

Ratio

Prob > F

5 2.21176 0.4423 31.761 <0.0001*

Error 18 0.25065 0.0139

Corr.

Total

23 2.4624

Figure 5. Percent Biofilm Inhibition of Varied Naringenin

Concentrations. Biofilm inhibition percentages

calculated as (control OD - treatment OD)/(control

OD)*100. Error bars show 1 Standard Error of the

Mean (SEM).

Figure 7. Percent Biofilm Inhibition of Varied Cinnamaldehyde

Concentrations. Shows the biofilm

inhibition percentages calculated as (control OD -

treatment OD)/(control OD)*100. Error bars show 1

SEM.

4.3 – Multiple Compounds Cellular Growth Assay

The experiment was conducted to show the growth rate

of the V. harveyi bacteria when treated with combinations

used in biofilm formation assays. Because the growth rates

22 | 2019-2020 | Broad Street Scientific BIOLOGY


of bacteria remained unchanged when treated with combinations

of compounds, it is understood that the treatments

are inhibiting only quorum sensing, not altering the mortality

of V. harveyi.

Figure 8. Cellular Growth Assay. Shows the growth

curve of V. harveyi treated with DMSO (control), cinnamaldehyde

(13.22 µg/mL)/naringenin (6.25 µg/mL),

neohesperidin (12.5 µg/mL)/naringenin (6.25 µg/mL)/

cinnamaldehyde (13.22 µg/mL), neohesperidin (12.5

µg/mL)/naringenin (6.25 µg/mL). Error bars show

mean +/- 1 SEM.

5. Discussion

The results of our study showed naringenin alone was

able to inhibit biofilm formation, whereas neohesperidin

and cinnamaldehyde alone at the concentrations tested

showed no consistent detectable effect on biofilm inhibition

(Fig. 2). This study confirmed previous studies showing

that naringenin strongly suppressed AI-2 and HAI-1

mediated quorum sensing in V. harveyi. Because cinnamaldehyde

was thought to inhibit the master regulatory protein

LuxR, it was hypothesized to be the most effective.

However, the variable effects of cinnamaldehyde as well

as the consistently undetectable effects of neohesperidin,

responsible for the inhibition of HAI-1 mediated quorum

sensing, were unexpected. Surprisingly, there was a

synergistic effect present when both cinnamaldehyde and

naringenin were used to inhibit biofilm formation. It is

important to note the combinations of cinnamaldehyde/

naringenin and neohesperidin/naringenin showed low

raw OD values. Only 1 mL of LM media was administered

to each well in later experiments to increase baseline

biofilm formation, so inhibition could more easily be detected.

In the earlier experiments, naringenin seemed to

overpower combinations as the combination was not significantly

different from naringenin alone. Therefore, half

concentrations of naringenin were tested with decreased

BIOLOGY

media per well. Synergy was detected with the full concentration

of cinnamaldehyde and naringenin.

One explanation for this synergy is that once naringenin

blocks AI-2 and HAI-1 mediated quorum sensing,

cinnamaldehyde is left to inhibit only the effects of CAI-1

mediated quorum sensing (Fig. 1). Cinnamaldehyde may

be more effective at decreasing the DNA-binding affinity

of relatively small amounts of LuxR, which would explain

the apparent synergy and why the synergy depended heavily

on the concentration of naringenin present. It is also

possible that there are more components to this pathway.

If naringenin inhibits other unknown components of this

pathway that cinnamaldehyde cannot, cinnamaldehyde

alone might be ineffective because of the activation of multiple

components of the pathway. The inconsistent results

of cinnamaldehyde could be explained by the presence

of these additional, confounding components. In combinations,

however, naringenin could block the unknown

components and cinnamaldehyde would be able to work

effectively. In addition, it is confirmed that these compounds

inhibit only quorum sensing and not cell growth

rates because results of the cellular growth assay show a

similar curve with and without the treatments (Fig. 8).

Future directions include altering the concentrations of

quenchers to see if greater inhibition and optimization are

possible, elucidating the exact mechanisms of these quorum

quenchers, and observing the effects of combinations

of different quorum quenchers on biofilm formation. It

would be interesting to see which autoinducers contribute

the greatest effect on biofilm formation by testing combinations

of naringenin or cinnamaldehyde with a quorum

quencher that acts on CAI-1 mediated quorum sensing.

Another approach would be to test knockout strains of

V. harveyi that only respond to AI-2 and HAI-1 mediated

quorum sensing to see if decreased biofilm formation

occurs. As further research ensues, the mechanisms of the

quorum quenchers will be better known and can be optimized

in combinations to produce more synergistic effects

on decreasing biofilm formation.

More broadly than decreasing biofilm formation, quorum

sensing inhibition has the potential to act as an alternative

to antibiotics. Because quorum quenching effectively

reduces biofilm formation and does not kill bacteria,

it can be a less selective alternative therapy to antibiotics

[12]. Since quorum sensing controls not only biofilm formation

but also the secretion of other virulence factors [2],

the inhibition of quorum sensing can decrease virulence

factor production. The production of these factors can also

be measured to detect synergistic effects among quorum

quenchers. The applications of this project are in inhibiting

biofilm formation, which can be used as the foundation for

biofilm-resistant hospital materials, alternative antibiotic

treatments or as a model for V. cholera. These applications

are important as a model for V. cholera because the V. cholera

biofilm plays a critical role in pathogenesis and disease

Broad Street Scientific | 2019-2020 | 23


transmission [13]. A cholera outbreak occurred in Haiti in

2010 when UN peacekeepers introduced the disease to the

earthquake devastated country, taking upwards of 10,000

lives. In 2018, there were cholera outbreaks in Yemen and

Somalia. Additionally, costs of outbreaks were estimated

to be $38.9-$64.2 million in 2005 [14]. Through further

studying this synergistic effect of quorum quenchers on

decreasing biofilm formation in V. harveyi, the mortality

rates and economic costs due to infectious diseases including

cholera can decrease in the future.

6. Acknowledgments

This work would not have been possible without the

endless guidance and support of my mentor, Amy Sheck.

I would like to extend gratitude to Kimberly Monahan, as

well. Also, I would like to thank Nick Koberstein for help

with statistical analysis. Furthermore, I would like to thank

the Research in Biology Classes of 2019 and 2020 for their

friendship and advice. Lastly, I am highly indebted to the

Glaxo Endowment to the North Carolina School of Science

and Mathematics for providing the funding for this

project.

7. References

[1] Bassler, B. L. & Losick, R. (2006) Bacterially speaking.

Cell, 125, 237-246.

[2] Defoirdt, T. (2017) Quorum-sensing systems as targets

for antivirulence therapy. Trends in Microbiology, 26, 313-

328.

[3] Vadakkan, K., Choudhury, A. A., Gunasekaran, R.,

Hemapriya, J. & Vijayanand, S. (2018) Quorum sensing intervened

bacterial signaling: pursuit of its cognizance and

repression. Journal of Genetic Engineering and Biotechnology,

16, 239-252.

[7] Lilley, B. N. & Bassler, B. L. (2000) Regulation of quorum

sensing in Vibrio harveyi by LuxO and Sigma-54. Molecular

Microbiology, 36, 940-954.

[8] Kalia, V. C., Patel, S. K. S., Kang, Y. C., & Lee, J. (2019)

Quorum sensing inhibitors as antipathogens: biotechnological

applications. Biotechnology Advances, 37, 68-90.

[9] Vikram, A., Jayaprakasha, G. K., Jesudhasan, P. R.,

Pillai, S. D. & Patil, B. S. (2010) Suppression of bacterial

cell-cell signaling, biofilm formation and type III secretion

system by citrus flavonoids. Journal of Applied Microbiology,

109, 515-527.

[10] Packiavathy, I. A. S. V., Sasikumar, P., Pandian, S. K.,

& Ravi, A. V. (2013) Prevention of quorum-sensing-mediated

biofilm development and virulence factors production

in Vibrio spp. by curcumin. Applied Microbiology and Biotechnology,

97, 10177-10187.

[11] Ta, C. A. K. & Arnason, J. T. 2015. Mini review of

phytochemicals and plant taxa with activity as microbial

biofilm and quorum sensing inhibitors. Molecules, 21, E29.

[12] Kalia, V.C. (2013) Quorum sensing inhibitors: an

overview. Biotechnology Advances, 31, 224-245.

[13] Silva, A. J. & Benitez, J. A. (2016) Vibrio cholera Biofilms

and Cholera Pathogenesis. PLOS Neglected Tropical

Diseases, 10(2), e0004330.

[14] Tembo, T., Simuyandi, M., Chiyenu, K., Sharma, A.,

Chilyabanyama, O. N., Mbwili-Muleya, C., Mazaba, M. L.,

& Chilengi, R. (2019) Evaluating the costs of cholera illness

and cost-effectiveness of a single dose oral vaccination

campaign in Lusaka, Zambia. PLOS One, 14(5), e0215972.

[4] Anetzberger, C., Pirch, T. & Jung, K. (2009) Heterogeneity

in quorum sensing-regulated bioluminescence of

Vibrio harveyi. Molecular Microbiology, 73(2), 267-277.

[5] Tu, K. C. & Bassler, B. L. (2007) Multiple small RNAs

act additively to integrate sensory information and control

quorum sensing in Vibrio harveyi. Genes Dev, 21, 221-223.

[6] Brackman, G., Deifoirdt, T., Miyamoto, C., Bossier,

P., Calenbergh, S. V., Nelis, H., & Coenye, T. (2008) Cinnamaldehyde

and cinnamaldehyde derivatives reduce virulence

in Vibrio spp. by decreasing the DNA-binding activity

of the quorum sensing response regulator LuxR. BMC

Microbiology, 8, 149-163.

24 | 2019-2020 | Broad Street Scientific BIOLOGY


EMOTIONAL PROCESSING AND WORKING MEMORY

IN SCHIZOPHRENIA WITH NEUTRAL AND NEGATIVE

STIMULI: AN fMRI STUDY

Cindy Zhu

Abstract

Schizophrenia (SZ) is a psychiatric disorder that results in abnormalities with emotional processing and working memory.

Working memory (WM) and Emotional Processing (EP) are supported by specific neural regions in the brain; however,

we do not know how these regions interact to influence task performance. To explore the effects of emotional valence

on working memory, an emotional n-back task was used with a focus on Neutral and Negative emotional valence and

on working memory and emotional processing regions in the brain. In the neutral condition, we observed recent-onset

schizophrenia (RO) participants having a lower activation than genetic high-risk (HR) participants in working memory

regions (ACC, Left ACC, Right ACC), and in ventral regions (Left NAcc). RO showed lower activation than control

(CON) participants in the Left DLPFC. These results align with behavioral results - RO did not show a significant difference

in D′ values between Neutral and Negative conditions while controls did, which may reflect the belief that RO

patients have a greater ability to focus on the working memory task due to tunneling and focus less on the emotional valence

as compared to the control patients. When exploring the correlation between brain activity and task performance,

we found CON and RO exhibiting a positive correlation, while HR showed a negative correlation. In conclusion, during

a cognitively demanding task, RO participants exhibit fewer differences in working memory and emotional regulation

between conditions, which supports prevailing theories of emotional blunting and tunneling. In addition, surprising

differences between HR and RO groups showed HR with overactivation, as opposed to prevailing theories that HR will

show levels of activation intermediate between CON and RO.

1. Introduction

Schizophrenia is characterized by significant impairments

in working memory and emotional regulation.

Impairments in working memory (WM), the temporary

storage and manipulation of information held for a limited

period of time, are considered a core neurocognitive deficit

in patients with schizophrenia (SZ) and can significantly

impact quality of life. Deficits in emotional regulation, or

the conscious effort in modulating emotional response to

goal-unrelated or irrelevant emotional stimuli, interfere

with social life and daily functioning, and may also be associated

with the development of psychotic symptoms

[1]. These deficits in WM are present prior to the onset

of illness, such as in genetically high-risk individuals, or

in patients who are medication-free in their first episode

of illness [2]. Thus, gaining a better understanding of the

interplay of emotional regulation and working memory in

patients with schizophrenia may help improve patient diagnosis

and quality of life.

Emotional blunting, or blunted affect, is one of the core

negative symptoms of schizophrenia and results in difficulty

in expressing emotions in reaction to emotional stimuli

due to problems with emotional processing. There are

mixed findings of activation during emotionally evocative

stimuli between people with and without schizophrenia in

areas associated with emotion, with some studies reporting

no differences in activation, while others report diminished

activation [3]. Emotional processing abnormalities

in schizophrenia have been shown to reduce activation

in brain regions associated with emotional processing in

response to emotionally evocative stimuli [4]. It remains

unclear how these alterations in emotional valence processing

may impact WM functions for individuals with

schizophrenia.

Working memory is an important component of higher

cognition, such as goal-directed behavior. Deficits in

schizophrenia may relate to a number of other core symptoms

in schizophrenia [5]. Working memory impairment

is often associated with differing DLPFC activation, which

is implicated in executive functions, goal-directed planning,

and inhibition, and is a part of working memory circuitry

[6]. The effects of these deficits can be studied using

n-back tasks, which test working memory by utilizing continuous

updating and order memory [2].

Emotional regulation processes and working memory

can be studied in conjunction using emotional working

memory paradigms through functional magnetic resonance

imaging (fMRI). Task-based fMRI allows for both

behavioral and aggregate neural activation measurements

and can help identify how emotion and WM interact. An

emotional one-back task was used with neutral and negative

conditions. Participants were instructed to indicate

when a new stimulus (image with emotional valence)

was the same as the stimulus presented one before, which

required short-term memory engagement. By using an

BIOLOGY

Broad Street Scientific | 2019-2020 | 25


emotional one-back task, the influence of the emotional

valence (i.e. whether a stimulus is neutral or negative) on

WM can be calculated.

The influence of emotion on cognition is an essential

topic to research, but research on this topic has received

less attention than others in schizophrenia literature. In

controls, it has been shown that emotional stimuli can

garner more attention than neutral stimuli. This extra attention

may facilitate the processing of emotional stimuli.

In patients with schizophrenia, there may be reduced activation

and worsened performance due to the addition

of emotional valence to the WM task. DLPFC has an important

role in the integration of emotional and cognitive

information [7]. Studies of negative valence and cognition

tend to produce more robust results due to the arousing

nature of the images used. Other studies [4] also explore

the interaction between EP and WM, but only with control

and schizophrenia patient groups. By including HR

participants, which have been found to show similar

working memory deficits as those with schizophrenia [2],

in our study of the interaction of WM and EP, we address a

new set of questions in how high-risk participants perform

compared to controls and RO and if they exhibit deficits

similar to RO participants.

The overarching goal of this study is to identify how

WM and emotional processes interact, particularly in the

context of schizophrenia. In order to achieve this aim, we

administered an emotional n-back task to 76 participants

(35 control, 20 HR, 21 RO) to determine group differences

in regional activation and WM performance associated

with psychotic illness. We hypothesized that RO participants

would perform consistently across neutral and negative

conditions (because emotional blunting may result in

less expressed performance differences between valences),

while control participants would have more impaired performance

in the negative condition compared to the neutral

condition. Moreover, we hypothesized that changes in

performance between conditions will differ between subject

groups. When looking at brain activation, we hypothesized

that the control participants would exhibit a greater

change in ventral regions when comparing between the

neutral and negative conditions and that control participants

would have a greater change in activation in WM

regions between neutral to negative. We also hypothesized

that genetic high-risk participants would exhibit brain activation

and task performance intermediate between CON

and RO. Finally, we investigated links between activation

and behavior to see if changes in activation correlated with

changes in task performance.

2. Methods

2.1 – Participants

Twenty-one patients with recent-onset schizophrenia

and twenty genetic high-risk patients were recruited from

the UNC Healthcare System. Thirty-five healthy control

subjects were also included. All participants provided

written consent to the study approved by the University

of North Carolina- Chapel Hill IRB. All participants were

between the ages of 16-45, of any ethnicities or gender,

had no presence of metallic implants or devices interfering

with MRI, and were not pregnant. Inclusion criteria for

recent-onset schizophrenia (RO) patients were: (1) Meet

DSM-IV criteria for SZ or schizophreniform disorder, (2)

No history of major central nervous system disorder or intellectual

disability (IQ<65), (3) Must have illness for <5

years, (4) No current diagnosis of substance dependence,

and no substance abuse for 6 weeks. RO patients were also

instructed to refrain from taking benzodiazepine medications

on the morning of testing but instead to bring their

medication with them to take after scanning. Inclusion

criteria for genetic high risk (HR) patients were: (1) Must

have first degree relative with psychotic disorder, (2) Must

not meet DSM-IV criteria for past or current Axis I psychotic

disorder on bipolar affective disorder, (3) No history

of major central nervous system disorder or intellectual

disability (IQ<65), (4) No current treatment with antipsychotic

medication. Healthy controls (CON) were excluded

if they had history of a DSM-IV axis I psychiatric disorder,

family history of psychosis, history of current substance

abuse/dependence, history or current medical illness that

could affect brain morphology, or clinically significant

neurological or medical problems that could influence the

diagnosis or the assessment of the biological variables in

the study. All participants gave written informed consent

consistent with the IRB of UNC if over 18 or assent and

parent/guardian provided consent for minors prior to

their participation in the study.

2.2 – Emotional One-back Task and Procedure

Each participant completed an emotional one-back task

with an auditory component during a functional magnetic

imaging (fMRI) session with 8 runs. The emotional oneback

task consists of a visual tracking task, using images

with either Positive, Neutral, or Negative valence, as defined

by the International Affective Picture System (IAPS)

[8]. Further analysis was performed with only Neutral and

Negative Valences. Patients with schizophrenia report

feeling negative emotion strongly but are less outwardly

expressive of this negative emotion [9]. By focusing on the

Negative valence, we want to observe if RO exhibit similar

activation to CON during negative emotional situations,

as the contexts in which RO patients experience negative

emotion is different than those without SZ. All images in

the run were of the same valence, and subjects were asked

to press a button when they saw the same image two times

in a row (Fig. 1). A control condition with no n-back

task was also included. The auditory component, which

occurred simultaneously with the visual component, involved

subjects hearing irrelevant standard and pitch devi-

26 | 2019-2020 | Broad Street Scientific BIOLOGY


ant tones at random intervals, but for the study’s purposes,

it was excluded from further analysis.

The task was divided into 8 runs, with 2 runs of each

valence, and 2 control runs with no n-back task. Each

run lasted 200.58s, with 14 target images, 56 non-target

images. Each image was presented for 500ms and the inter-stimulus

interval was either 1500ms or 3000ms.

Figure 1. Emotional One-Back Task. Participants

were asked to press a button when the picture shown

matches the picture shown one before.

2.3 – Behavioral Analysis

To analyze the performance of participants during the

task, D′, or sensitivity index, was used as a metric, because

it considers accuracy and sensitivity. Paired t-tests were

performed to investigate within-group differences in D′

between Neutral and Negative task conditions. To investigate

differences between groups (CON, HR, and RO)

and condition (Neu and Neg) and the interaction effects

of both factors on D′, a 2-way ANOVA was used. Finally,

a one-way ANOVA was used to investigate group differences

for neutral and negative conditions, only considering

one condition at a time. A Tukey comparison of means

was performed for One-way and Two-way ANOVA’s that

had significant results to further interpret the results.

2.4 - Neuroimaging Analysis

2.4.1 - Imaging Data Acquisition

A General Electric 3.0 T MRI scanner with a functional

gradient-echo echo-planar imaging sequence allowing

for full-brain coverage (TR: 2000 ms; TE: 27 ms; FOV:

24 cm; image matrix: 64×64; flip angle: 60; voxel size:

3.75×3.75×3.8mm 3 ; 34 axial slices) was used. Each functional

run had 120 time-points. Structural MRIs were acquired

before fMRIs to obtain 3D coplanar anatomical T1

images using a spoiled gradient-recalled acquisition pulse

sequence (TR: 5.16 ms; TE: 2.04 ms; FOV: 24 cm; image

matrix: 256×256; flip angle: 20; voxel size: 0.94×0.94×1.9

mm 3 ; 68 axial slices).

2.4.2 - Preprocessing

Functional data analyses were carried out using the

University of Oxford’s Center for Functional Magnetic

Resonance Imaging of the Brain (FMRIB) Software Library

(FSL) release 6.0 [11]. Image preprocessing consisted

BIOLOGY

of using the Brain Extraction Tool (BET) to remove nonbrain

structures, motion correction, time-slice correction

and spatial filtering using a Gaussian kernel of full width

half maximum 5 mm, high-pass temporal filtering. The

functional images were co-registered to the structural images

in their native space. These images were then normalized

to the Montreal Neurological Institute standard

brain. The default adult template in FMRIB’s Linear Image

Registration Tool (FLIRT) was used for registration [10].

FMRIB’s Improved Linear Model (FILM) was used for

pre-whitening in order to estimate and account for each

voxel’s time series auto-correlation.

2.4.3 - Whole Brain Analysis

Whole Brain Analysis was performed to ensure that

the tasks activated regions known to be involved in WM

and EP. These whole-brain maps were thresholded for significance

to obtain a significant rate of p=0.05 (corrected

for false discovery rate due to multiple comparisons) with

Randomise, which tests for correlation inferences using

permutation methods, implemented in FSL [11]. The Dorsal

Attention and Ventral Attention networks [12] were

used to check that existing patterns of activation in control

participants were consistent with working memory and

emotional processing regions. The control mean group

activation was then overlaid on top of the two attention

networks to see where activation overlapped.

2.4.4 - Region of Interest Selection

Regions of Interest (ROI) involved in working memory

circuitry and/or emotional processing were selected

(Fig. 2). Dorsal, or working memory regions, include the

whole anterior cingulate cortex (ACC), left and right anterior

cingulate cortex (Left ACC, Right ACC), left and right

dorsal lateral prefrontal cortex (Left DLPFC, Right DLP-

FC), and the interparietal sulcus (IPS). Ventral regions,

or regions involved in emotional processing, include left

and right amygdala (Left AMY, Right AMY), left and right

nucleus accumbens (Left NAcc, Right NAcc), and left and

right Orbitofrontal cortex (Left OFC, Right OFC).

2.4.5 - fMRI Statistics

Using FSL version 6 [11], the FEAT first-level analysis

was performed on each run per subject, where a general

linear model is used to model the expected hemodynamic

response function, which models the evoked hemodynamic

response to a neural event. In the second-level analysis,

three contrast of interest (Neutral, Negative, Neutral

> Negative) were generated for each subject across multiple

runs to investigate the interaction between emotional

valence and type of model design (NeuVisTarg > NegVis-

Targ or NeuVis > NegVis), and each emotional valence

and type of model design (NeuVisTarg and NegVis). In the

final analysis, only event designs Neutral, Negative, and

Neutral>Negative contrasts were considered.

Broad Street Scientific | 2019-2020 | 27


Table 1: Demographic information for the participants. CON denotes Healthy controls, HR denotes High Risk

Participants and RO denotes Recent Onset Participants.

CON (n=35) HR (n = 20) RO (n = 21)

Mean SD/% Mean SD/% Mean SD/% p(CON-HR) p(CON-RO) p(HR-RO)

Age 27.01 6.07 30.62 7.89 25.29 4.92 0.068 0.284 0.015

Male 19 54.29% 5 25% 17 80.95% 0.037 0.0457 0.0004

Female 16 45.71% 15 75% 4 19.05% 0.037 0.0457 0.0004

These contrasts were then submitted to the third level

analysis to examine differences between groups in task-related

brain activity. Region of Interest analysis was then

conducted using FEATQUERY, and beta values, which

represent the direction and magnitude of regional brain

activation, were extracted from working memory (WM)

and emotional processing (EP) of the brain.

We conducted paired t-tests to investigate the differences

between neutral and negative conditions within the

same subject group. To investigate differences between

groups (CON, HR, and RO) and condition (Neu and

Neg) and the interaction effects of both factors, a 2-way

ANOVA was used. Finally, a one-way ANOVA was used

to investigate group differences for neutral and negative

conditions for different regions. Post Hoc analysis was performed

with a Tukey comparison of means for one way

and Two-way ANOVA’s that had significant results, in order

to isolate specific effects (or differences).

3. Results

3.1 - Basic Information

The demographic information of these three groups

(Healthy Controls, High-Risk participants, and Recent

Onset Participants) are listed in Table 1. There was no significant

difference between CON and RO regarding age.

However, there is a significant age difference between

CON-HR and HR-RO, which may be due to the greater

ages of the HR participants. High-Risk participants may

be of greater age because the tendency for schizophrenia

to manifest in patients in their late teens or early twenties

does not restrict the pool of HR participants. As a result,

there is a wider age range for HR participants to be chosen

from (unrestricted by diagnosis), while recent-onset

schizophrenia typically occurs before the age of 35, which

may explain the difference in mean age. There is a significant

difference between all groups (CON-HR, CON-RO,

RO-HR) regarding sex (male and female). This is due to

the high percentage of Male RO and a high percentage of

Female HR.

3.2 - Behavioral Performance on the Emotional One-back

Task

3.2.1- Within-Group Analysis

Control participants performed significantly worse

during the neutral valence than the negative valence (t

= -2.216, p = 0.035) but the recent-onset and high-risk

participants exhibited no significant difference in performance

between the neutral and negative valences (Fig. 3).

Figure 2. Regions of Interest used in Analysis.

2.5 - Brain Activity and WM Performance

In order to explore links between brain activity and

WM performance, linear regression was performed

with extracted beta values from selected ROI and neutral

and negative conditions as the independent variable and

D′ values as the dependent variable. Subjects without

behavioral data were excluded. These regressions were

performed on patient groups (CON, RO and HR participants)

and the correlation was identified.

Figure 3. Within Group D′ Values for Neutral and

Negative Condition.

28 | 2019-2020 | Broad Street Scientific BIOLOGY


3.2.2 - Between-Group Analysis

A Two-way ANOVA with D′ as the dependent variable

revealed no significant main effects, including Subject

Group, Valence, and Subject Group by valence interaction.

While not significant, the HR group had the best task performance

(highest group mean D′), followed by the control

group and lastly the RO group.

3.3 - Whole Brain Imaging Analysis

For Neutral and Negative conditions, the activation of

the control means overlapped with dorsal and ventral regions.

Some specific regions include the anterior cingulate

cortex and the intraparietal sulcus. Neutral>Negative exhibited

less overlap, but since this condition is calculated

from the difference between Neutral and Negative, it is

expected that there would be lower activation in those regions.

Activation in visual regions is expected given that it

is a visual task (i.e. the occipital activation).

3.4 - Region of Interest Analysis

3.4.1 - Within-Group Analysis

CON, HR and RO subjects did not show significant

within-group changes in activation between Neutral and

Negative valence in Dorsal or Ventral Regions (Fig. 5).

While not significantly different, in control participants,

all regions but the left AMY had greater activation in the

Neutral condition. In high-risk participants, all dorsal regions

but the left and right DLPFC had greater activation

in the Neutral condition, and in ventral regions, all but

the left AMY and left NAcc had greater activation in the

Negative condition. Finally, in recent-onset participants,

all dorsal regions but the right ACC and right DLPFC had

greater activation in the Negative condition and all ventral

regions but the right AMY had greater activation in the

Negative condition.

Figure 4. Whole Brain Results showing Mean Control

Activation across conditions.

BIOLOGY

a) Neutral Condition b) Negative Condition c) Neutral > Negative Condition

Figure 5. Within-Group Average Beta Values for

Neutral and Negative Condition.

3.4.2 - Between-Group Analysis

We observed significant main effects of group in working

memory regions, specifically ACC (F=4.158, p=0.018),

Left ACC (F=3.434, p=0.0348), and Right ACC (F=4.3281,

p=0.015). Upon further investigation, the group difference

lies between RO and HR groups in the Neutral condition,

with greater activation in the HR group (Table 2). We also

observed significant main effects in emotional processing

regions, specifically, Left NAcc (F = 4.4413, p= 0.013).

Similarly in the Left NAcc, the group difference lies between

RO and HR groups in the Neutral condition, with

greater activation in the HR group (Table 2). As a follow-up

analysis of the Two-way ANOVA, the One-way

ANOVA was used to identify which valence the significant

differences found with the Two-way ANOVA were associated

with. There was a significant difference between

groups in the Left DLPFC during the Neutral condition

(F=4.0098, p=0.02227). After performing post-hoc analysis,

the difference was shown to be between the RO and

CON groups, with RO having significantly less activation

than CON (Fig. 6). The examination of beta values suggests

differences in the Neutral condition may be driven

Broad Street Scientific | 2019-2020 | 29


Table 2. Significant 2-way ANOVA and Post-Hoc Test Results

2-way ANOVA

Region F value p-value Group Sig. Diff Tukey Comparison of Means p-value Greater Activation

ACC 4.158 0.018 HR-RO 0.013 HR

Left ACC 3.438 0.035 HR-RO 0.027 HR

Right ACC 4.328 0.015 HR-RO 0.011 HR

Left NAcc 4.4413 0.013 HR-RO 0.009 HR

by greater deactivation or less activation for RO subjects in

dorsal and ventral regions. While not significant, HR and

Control activation is greater than RO for all regions in the

Neutral condition. HR shows the greatest activation across

all regions except the Left DLPFC (control greater), while

controls show greater activation than RO in all regions.

In the Negative condition, there are no significant differences

between groups. HR shows greater activation in

all regions, while controls show greater activation than

RO in all regions but the Left DLPFC. Finally, to assess

the change in activation from the Neutral condition to

the Control condition, Neutral > Negative activation was

evaluated. While there were no significant differences between

groups, in dorsal regions except for the right DLP-

FC and the IPS, RO exhibited greater deactivation. This

deactivation can be attributed to RO having greater activation

in Negative conditions than Neutral conditions for

those regions.

3.5 - Brain Activity and Task Performance

We explored the impact of brain activity in working

memory and emotional processing regions on participants’

performance by the subject group during Neutral

and Negative conditions. For control participants, there

was a significant link between activation and task performance

in the Right NAcc in the Neutral condition.

This linear relationship (t=2.125, p=0.0429) is positively

correlated, with greater activation linked with a higher

D′ (Fig. 7). High-Risk participants showed a significant

negative correlation between activation and task performance

in the Left ACC (t=-2.312, p=0.034) and Left NAcc

(t=-3.086, p=0.007) for the Neutral condition and in the

Left NAcc (t=-2.876, p=0.011) , Right NAcc (t=-3.294,

p=0.005) , and Right Orbitofrontal Cortex (t=-2.486,

p=0.024) for the Negative condition. As presented in Figure

7, when brain activation increased for High-Risk participants

in the regions above, their D′ decreased, showing

worse task performance with increased activation. Recent

Onset participants exhibited a significant positive correlation

between brain activity and task performance in

the Left ACC (t=2.303, p=0.040) and Left Orbitofrontal

Cortex (t=2.225, p=0.046) during the Negative condition.

Both Control and Recent Onset participants showed a

strong positive correlation between brain activity and

task performance, while High-Risk participants showed a

strong negative correlation.

a) Neutral Condition

b) Negative Condition

c) Neutral > Negative Condition

Figure 6. Between Group Average Beta Values for

different conditions.

30 | 2019-2020 | Broad Street Scientific BIOLOGY


4. Discussion

The behavioral and fMRI findings provide insights

into working memory and emotional regulation deficits

in patients with genetic high risk and recent-onset

schizophrenia. Our behavioral results show that patients

with RO Schizophrenia and participants in the HR group

did not exhibit a significant change in task performance

between the Neutral and Negative conditions, while the

control participants did. Imaging results of controls are

consistent with other studies and show activation in Dorsal

and Ventral regions. These results support the belief

that the dorsal and ventral regions chosen for this study

are activated during WM and emotional regulation tasks.

Our results may show the effects of emotional blunting

in RO participants, since in the Neutral > Negative condition,

for all ventral regions, RO did not exhibit significant

differences. Emotional blunting may cause this, as RO

participants are less impacted by the valence of the task

and redirect capacity to other functions, such as working

memory. This is reflected in their performance results as

well, with the least change in D′ from Neutral to Negative,

even if overall performance is worse. Increased activation

in WM regions was linked with better performance in

participants. The D′ values reflect this, as HR consistently

had higher activation than Controls and RO and also

performed better in the task.

When looking at the activation in the Neutral condition

between groups, we observed significant abnormal

deactivation/lower activation in ACC, Left ACC, Right

ACC, Left DLPFC and Left NAcc. In previous studies,

schizophrenic patients often have diminished activation

when compared to controls [3]. In looking at ventral

regions, the underactivation of the Amygdala in patients

with schizophrenia in response to negative stimuli may

not be due to underactivation in the Negative condition,

but due to overactivation in response to neutral stimuli

[3]. These findings reflect the emotional experience

reported by people with schizophrenia, where they often

experience more negative emotions in response to neutral

stimuli [3]. Previous studies show that other brain

regions also showed less activation, such as the ACC and

DLPFC [3], which our results reflect.

We also observed that HR participants showed consistent

overactivation in dorsal and ventral regions compared

to control participants for both the Neutral and Negative

conditions, which was surprising. Based on previous studies,

we expected HR to show levels of cognitive performance

intermediate between healthy controls and schizophrenic

patients [13]. Even though HR activation tended

to be more similar to control activation in scale, it was always

greater than Control and RO activation. One possible

interpretation is that HR participants are suffering from

WM and emotional processing deficits, so the activation is

greater in order to compensate for these deficits; RO may

just have less activation overall due to task tunneling and

having a lower capacity to spend on both WM and emo-

Figure 7. Significant links between D′ and Activation.

tional regulation.

Interestingly, our results suggest that during an effortful

cognitive task, less brain activity was associated with a significant

reduction in CON and RO patients’ performance

and more brain activity was associated with a significant

reduction in HR participants’ performance. HR exhibited

the opposite trend compared to RO and Control participants

when examining brain activation and task performance.

Further connectivity analysis could help us gain a

better understanding of why this occurs.

This study has several limitations. First, the sample size

of the participant groups (Con, HR, RO) was relatively

small and there were significant differences regarding participant

age and gender between groups, which may limit

the interpretability of the results. Secondly, only Neutral,

Negative, and Neutral > Negative conditions of the task

were considered. By utilizing the positive conditions and

other brain circuits, we might be able to provide more information

about the differences between groups during an

emotional one-back task. There was no co-variate included

for medication in the RO group, which may affect our

results. Finally, the behavioral results were only for a subset

of participants, which may limit the interpretability of

the D′ values and the linkage with performance.

In the future, we hope to utilize these results in patient

diagnosis and patient outcome. We also hope to explore

the interplay of symptoms of schizophrenia with WM and

emotional regulation and include different neural circuits

based on the regions involved.

BIOLOGY

Broad Street Scientific | 2019-2020 | 31


5. Conclusion

In summary, by using an emotional one-back task

allowing for the analysis of the interplay of WM and ER,

we demonstrated between-group differences attributed

to emotional blunting and surprising results for highrisk

participants. Our findings indicate that RO patients

tend to show deactivation/lesser activation in WM and

ER regions, which can be attributed to limited capacity

(and less activation overall), with fewer changes due

to emotional valence. Interestingly, our results suggest

that during an effortful cognitive memory task with no

emotional valence, HR does not fall between CON and SZ

as an intermediate, but instead exhibits greater activation.

Another surprising finding of HR was that less brain

activity was associated with a significant reduction in

CON and RO patients’ performance, while more brain

activity was associated with a significant reduction in HR

participants’ performance. Understanding the WM and

emotional processing deficits in schizophrenia is a critical

target for improving diagnosis and recovery outcomes in

schizophrenia.

6. Acknowledgments

I would like to acknowledge Dr. Andrea Pelletier-Baldelli,

Dr. Aysenil Belger and Mr. Josh Bizzell from the UNC

Department of Psychiatry for their help with understanding

the cognitive side of this project, as well as providing the

data used here, Mr. Robert Gotwals for his guidance during

the research process, and the Research in Computational

Sciences Program at NCSSM.

7. References

[1] Guimond, S., Padani, S., Lutz, O., Eack, S., Thermenos,

H., & Keshavan, M. (2018). Impaired regulation of emotional

distractors during working memory load in schizophrenia.

Journal of psychiatric research, 101, 14–20.

[2] Seidman, L. J., Thermenos, H. W., Poldrack, R. A.,

Peace, N. K., Koch, J. K., Faraone, S. V., & Tsuang, M. T.

(2006). Altered brain activation in dorsolateral prefrontal

cortex in adolescents and young adults at genetic risk for

schizophrenia: An fmri study of working memory. Schizophrenia

research, 85(1-3), 58–72.

[5] Forbes, N., Carrick, L., McIntosh, A., & Lawrie, S.

(2009). Working memory in schizophrenia: A meta-analysis.

Psychological medicine, 39(6), 889–905.

[6] Eryilmaz, H., Tanner, A. S., et al. (2016). Disrupted

Working Memory Circuitry in Schizophrenia: Disentangling

fMRI Markers of Core Pathology vs Other Aspects of

Impaired Performance. Neuropsychopharmacology, 41(9),

2411–2420. doi:10.1038/npp.2016.55

[7] Gray, J. R., Braver, T. S., & Raichle, M. E. (2002). Integration

of emotion and cognition in the lateral prefrontal

cortex. Proceedings of the National Academy of Sciences,

99(6), 4115–4120.

[8] Lang, P. J., Bradley, M. M., Cuthbert, B. N., et al. (1999).

International affective picture system (iaps): Instruction

manual and affective ratings. The Center for Research in

Psychophysiology, University of Florida.

[9] Kring, A. M. [A. M.], & Moran, E. K. (2008). Emotional

Response Deficits in Schizophrenia: Insights From

Affective Science. Schizophrenia Bulletin, 34(5), 819–834.

doi:10.1093/schbul/sbn071

[10] Jenkinson, M., Bannister, P., Brady, M., & Smith, S.

(2002). Improved optimization for the robust and accurate

linear registration and motion correction of brain images.

Neuroimage, 17(2), 825–841.

[11] Smith, S. M., Jenkinson, M., et al. (2004). Advances in

functional and structural mr image analysis and implementation

as fsl. Neuroimage, 23, S208–S219.

[12] Thomas Yeo, B., Krienen, F. M., et al. (2011). The organization

of the human cerebral cortex estimated by intrinsic

functional connectivity. Journal of neurophysiology, 106(3),

1125–1165.

[13] Bang, M., Kim, K. R., Song, Y. Y., Baek, S., Lee, E., &

An, S. K. (2015). Neurocognitive impairments in individuals

at ultra-high risk for psychosis: Who will really convert?

Australian & New Zealand Journal of Psychiatry, 49(5),

462–470.

[3] Kring, A. M. [Ann M], & Elis, O. (2013). Emotion deficits

in people with schizophrenia. Annual review of clinical

psychology, 9, 409–433.

[4] Becerril, K., & Barch, D. (2011). Influence of Emotional

Processing on Working Memory in Schizophrenia. Schizophrenia

Bulletin, 37(5), 1027–1038. doi:10.1093/schbul/

sbq009

32 | 2019-2020 | Broad Street Scientific BIOLOGY


SESAMOL AS A NOVEL REDOX MEDIATOR FOR

THE ELECTROCHEMICAL SEPARATION OF CARBON

DIOXIDE FROM FLUE GAS

Madison Houck

Abstract

The increasing presence of carbon dioxide in the atmosphere has contributed to overall rising temperatures over the past

several years. In particular, flue gas, which contains a mixture of carbon dioxide, nitrogen, and oxygen, is a common way

that large quantities of carbon dioxide are introduced to the atmosphere. Previously, the redox couple hydroquinone and

benzoquinone were applied to a fuel cell capable of separating carbon dioxide from other gases, but the negative environmental

impacts of these chemicals have prompted a search for an environmentally friendly option. This project seeks to

apply the proton-coupled electron transfer (PCET) reaction of sesamol to achieve the same effect. Cyclic voltammetry was

used to evaluate the electrochemical response of sesamol. Cyclic voltammograms show that sesamol undergoes a quasi-reversible

reaction in sodium bicarbonate, with peaks that correspond to a redox couple appearing after one cycle or after a

deposition. Additionally, half-cell liquid phase testing was performed, confirming that sesamol’s redox reaction creates a

pH gradient that drives carbon dioxide to be released at the anode. Furthermore, the construction of a fuel cell reveals that

with applied voltage, carbon dioxide concentration increases on the permeate side of the cell when sesamol is utilized as a

redox mediator. Future work can be done to further evaluate the efficiency of a sesamol fuel cell as compared to a quinone

fuel cell, and to confirm that the system selectively transports carbon dioxide, and not other components of flue gas.

1. Introduction

As global temperatures continue to rise as a result of

carbon dioxide’s increasing presence in the atmosphere,

the wide-reaching impacts of climate change have driven

researchers to search for new carbon dioxide remediation

techniques [1]. Many have been developed, including the

use of metal-organic frameworks, the direct reduction of

carbon dioxide, and the use of nucleophiles to bind to carbon

dioxide [2]. However, some of these techniques fall

short when it comes to practical applications because of the

large amounts of carbon dioxide that enter the atmosphere

through flue gas. Flue gas is created when fuel or coal is

burned, and is composed of oxygen, nitrogen, water vapor,

carbon dioxide, and trace amounts of other gases. Many of

the carbon dioxide capture technologies mentioned above

are rendered useless when exposed to flue gas; oftentimes,

water or nitrogen gas interfere with the mechanisms used.

Some technologies are simply not selective for carbon dioxide,

and some are structurally damaged by the presence

of other gases [3].

This problem prompted research into an applicable

technology that uses a fuel cell to electrochemically separate

carbon dioxide from other gases. These technologies

rely on carbon dioxide’s unique sensitivity to pH changes.

A pH gradient is established in the fuel cell that captures

carbon dioxide, and has no impact on the gases that might

be present in the mixture. Such a gradient is established

through the use of a redox mediator that undergoes a reversible

proton-coupled electron transfer reaction. The

Figure 1. Schematic for a fuel cell capable of separating

carbon dioxide from flue gas electrochemically

by establishing a pH gradient.

most widely studied redox mediator is a quinone redox

couple (shown in Figure 1). Quinones undergo a reversible

redox reaction. The reversibility of the reaction is important

because it means that the fuel cell can be used repeatedly,

an important consideration when considering any

wide-scale application. Additionally, the redox reaction

that the quinone couple undergoes is a proton-coupled

electron transfer reaction (PCET). While electrons are being

transferred to oxidize or reduce a species, a proton is

also transferred. It is this transfer of protons that drives

the pH gradient. As voltage is applied to the fuel cell, the

quinone is reduced at the cathode and oxidized at the an-

CHEMISTRY

Broad Street Scientific | 2019-2020 | 33


ode. As it is reduced, protons are consumed, and thus the

local pH becomes basic. The basic pH leads to the capture

of carbon dioxide as a bicarbonate ion because this capture

proves sensitive to subtle pH changes. At the other end of

the cell, quinone is reduced, releasing protons, decreasing

the pH, and regenerating carbon dioxide gas from bicarbonate

ions, that then exits the cell. Carbon dioxide is the

only gas that exits; nitrogen, water, or oxygen are not impacted

by pH changes and do not cross the cell [3].

Unfortunately, although a quinone couple has been

demonstrated to work in a fuel cell to help transfer carbon

dioxide across a membrane, it is not a perfect solution.

Quinones have negative impacts on the environment and

enter largely as air pollutants; since the goal of the fuel cell

is environmental remediation, there has been a push to

move away from using quinones in practical applications

in this technology [4]. Additionally, a species in a common

quinone couple, hydroquinone, is a suspected carcinogen

[5], prompting a search for a more environmentally friendly

option that has similar redox behavior. One compound

that has been singled out for its quinone-like properties is

sesamol [4]. Isolated from sesame seeds, sesamol is unlikely

to have a negative impact on the environment. It is composed

of a fused ring structure, and has a hydroxyl group

attached to the benzene ring. Sesamol’s redox mechanism

is shown in Figure 2 and is slightly more complex than

that of a quinone; the reaction is quasi-reversible. The first

step of the reaction is irreversible, but it creates a quinone

structure (2-hydroxymethoxybenzoquinone or 2-HMBQ)

that then undergoes a second reduction reaction (to form

2-hydroxymethoxyhydroquinone or 2-HMHQ). This second

reaction is reversible, and largely mimics a quinone’s

redox behavior in that it transfers a proton as well [6].

Figure 2. Scheme proposed for the oxidation of sesamol

in aqueous solutions, forming 2-hydroxymethoxybenzoquinone

that undergoes further reduction.

This project aims to identify sesamol as a more environmentally

friendly alternative to quinones in a fuel cell

used for the separation of carbon dioxide from other gases.

First, sesamol’s quasi-reversible reaction was studied

through cyclic voltammetry in sodium bicarbonate and

saturated with both argon and carbon dioxide to examine

if conditions present in a fuel cell would fundamentally alter

the mechanism of the reaction. Two peaks corresponding

to the reversible redox reaction appeared upon repeated

sweeps, demonstrating that a reversible reaction begins

to occur after an irreversible step. Next, it was confirmed

that sesamol undergoes a PCET reaction; this was accomplished

through the use of half-cell liquid-phase testing,

which saw an increased current and gas evolution when

in sodium bicarbonate and saturated with carbon dioxide.

Finally, sesamol was used as a redox mediator in a fuel cell

and carbon dioxide transport across the cell was achieved.

2. Materials and Methods

2.1 – Materials

The polypropylene membrane, Celgard 3501 was a generous

gift from Celgard (Charlotte, NC). The Toray Carbon

Paper 060 electrode was purchased from The Fuel Cell

Store. Gases were industrial grade purchased from Airgas.

All other chemicals were purchased from Sigma-Aldrich,

and used without further purification.

2.2 – Cyclic Voltammetry

Cyclic voltammetry was performed on 10 mM 2,6-dimethylhydroquinone,

10mM 2,6-dimethylbenzoquinone,

and 1mM sesamol with 0.5 M sodium bicarbonate as the

analyte solution using an eDAQ potentiostat (ER466) and

a three electrode arrangement. The concentration of sesamol

was decreased in order to better examine each peak

that appeared. Each trial was performed in a 4 mL conical

vial. The reference electrode was silver/silver chloride, the

counter electrode was platinum/titanium, and the working

electrode was glassy carbon. The working electrode

was polished with a 0.3μM alumina suspension, and then

rinsed with acetone and water. The reference and counter

electrodes were rinsed with acetone and water. The

electrodes were all cleaned between each trial. Solutions

that were saturated with gas underwent 10 minutes of gas

sparging with argon gas, and then 10 more minutes of

sparging with carbon dioxide gas, as necessary. The pH of

each sample was measured with a Vernier pH Probe. Each

trial consisted of three sweeps in total, measuring the continued

electrochemical response of each sample. The scan

rate was 100mV/s for each trial, and data were collected

from -1.0V to 1.5V.

2.3 – Half Cell Liquid Phase Testing

Half-cell testing was performed in a 4 mL conical vial.

The working electrode was a 1cm by 3cm Toray Carbon

Paper 060 electrode (Fuel Cell Store), the reference electrode

was silver/silver chloride, and the counter electrode

was platinum. The carbon paper electrode was rinsed with

acetone and water between each trial, as were the counter

and reference electrodes. The catalyst was added from a

10mg/mL solution of 20% wt. platinum on carbon black

in methanol, which was drop-casted onto the carbon pa-

34 | 2019-2020 | Broad Street Scientific CHEMISTRY


per electrode at a concentration of 10μL/cm 2 and allowed

to evaporate. A voltage of 0.5V was applied using eDAQ

Chart software. Each solution was tested at a concentration

of 100mM in 0.5M sodium bicarbonate or 0.5M sodium

sulfate.

2.4 –Fuel Cell Construction

To construct the fuel cell, first a polypropylene membrane

(Celgard 3501) was soaked in a solution mixture of

5mM 2,6-DMBQ and 5mM 2,6-DMHQ or 5mM 2-HMBQ

and 5mM 2-HMHQ for 24 hours. Sesamol was oxidized

electrochemically by applying a voltage of 900mV for 60

seconds to the entire sample, then taking half of that sample

by volume and applying a voltage of 400mV for an additional

60 seconds [6]. Two carbon paper electrodes were

cut to 25cm 2 and placed on either side of the membrane,

with the catalyst layer facing the membrane. The catalyst

was added from 10uL/cm 2 of a 50mg/mL solution of 20%

wt. platinum on carbon black in methanol. The membrane

electrode assembly was applied to a fuel cell, and efficiency

was evaluated using fourier-transform infrared spectroscopy.

The permeate side of the fuel cell was attached to the

FTIR. Carbon dioxide was flowed into the fuel cell at a rate

of 5 standard cubic centimeters (SCCM) per minute, with

the sweep gas (argon) at a high flow rate, preventing the

CO 2

from crossing the membrane. Then, after 10 minutes,

once the fuel cell was saturated, the CO 2

was turned off and

the argon flow rate was decreased significantly. An FTIR

scan was taken with no additional voltage applied to measure

how much carbon dioxide was diffusing across the

membrane initially. Then, a voltage of 2.5 V was applied

with a voltmeter and FTIR scans were taken after 1, 5, and

10 minutes. The voltage was then turned off for 5 minutes

and another FTIR scan was taken to examine if CO 2

was

simply leaking across the fuel cell over time.

The redox reaction of 2,6-dimethylhydroquinone has

been well studied because of its reversibility [4]. It has a

clear oxidation peak (at 0V) and reduction peak (at -0.2V).

The reaction proceeds this way under argon and carbon

dioxide saturated conditions. Upon repeated scans, the

shape of the voltammogram does not change appreciably.

a)

b)

c)

3. Results

3.1 – Cyclic Voltammetry

Figure 4. Cyclic voltammograms for sesamol in 0.5M

acetic acid with (a) no gas saturation (oxidation steps

labelled), (b) argon gas saturation, and (c) carbon dioxide

gas saturation.

Figure 3. Cyclic voltammogram of 2,6-DMHQ in .5M

sodium bicarbonate.

Cyclic voltammetry was also used to characterize sesamol’s

redox reaction. The quasi-reversible scheme proposed

by Brito et al. is supported by the appearance of

secondary peaks referred to as “shoulder peaks” after one

voltage cycle (Fig. 4a). The oxidation peak that initially appears

around 0.7V represents the irreversible step that creates

2-HMBQ. After this species is present, the reduction

peak at 0V corresponds to the creation of 2-HMHQ which

CHEMISTRY

Broad Street Scientific | 2019-2020 | 35


is then oxidized. This oxidation peak, one of the shoulder

peaks, is noticeable in the next scan. This occurs around

0.4V. Additionally, after the shoulder peak appears, the

shape remains the same upon repeated scans, indicating a

continued and predictable electrochemical response. Under

argon saturated conditions, the voltammograms appear

similar. Additionally, when the sample was purged

with argon and then saturated with carbon dioxide, the

voltammogram shape remains consistent. Past a potential

of 1V, the aqueous solution begins to oxidize, thus the pertinent

data remain between -1V and 1V on the voltammogram.

a)

The redox mechanism of sesamol was next examined

in a solution of sodium bicarbonate instead of acetic acid

(Fig. 5). The sodium bicarbonate solution yields a much

higher pH (9.52 instead of 2.49), which could be important

to consider in the mechanism of sesamol oxidation, as it is

assumed that both protons and electrons are transferred.

There is less peak definition in the sodium bicarbonate;

however, despite this, shoulder peaks can still be observed,

albeit at slightly different potentials. Here, the irreversible

oxidation step likely occurs between 0.3V and 0.4V, and

the shoulder peak first appears around 0.1V. There is the

slight appearance of a reduction peak around -0.4V which

corresponds to the reduction of the 2-HMBQ. Under argon

and carbon dioxide conditions, the general shape of

the voltammogram is retained.

3.2 – Half Cell Liquid Phase Testing

Half-cell liquid phase testing was performed on samples

of hydroquinone and sesamol (Fig. 6). When potentials

were applied, the current was measured and the sample

b)

a) b)

100mM hydroquinone

solvent

saturated

with carbon

dioxide?

0.5 M no no

NaHCO 3

0.5 M yes yes

NaHCO 3

gas

evolved?

0.5 M no no

Na 2

SO 4

c)

Figure 5. Cyclic voltammograms of 1mM sesamol in

.5M sodium bicarbonate with (a) no gas saturation,

(b) argon gas saturation, and (c) carbon dioxide gas

saturation.

0.5 M yes no

Na 2

SO 4

c) d)

100mM sesamol

solvent

saturated

with carbon

dioxide?

0.5 M no no

NaHCO 3

0.5 M yes yes

NaHCO 3

0.5 M no no

Na 2

SO 4

0.5 M yes no

Na 2

SO 4

gas

evolved?

Figure 6. Data compiled from half-cell testing on (a)

100mM 2,6-DMHQ and (c) 100mM sesamol. Photograph

of carbon paper electrode after half-cell testing

was performed on (b) 2,6-DMHQ and (d) sesamol,

showing gas evolution.

36 | 2019-2020 | Broad Street Scientific CHEMISTRY


was observed for the evolution of a gas. When a gas is

evolved, it indicates that carbon dioxide is being released

at the anode, after the oxidation of the redox mediator and

the subsequent release of protons, driving down the local

pH. At a lower local pH, carbon dioxide is released from

a bicarbonate ion back into its gaseous form and exits the

solution. The presence of a gas therefore indicates that

the reaction transfers protons as well as electrons; if only

electrons were transferred, the pH would not change and

the carbon dioxide would remain in solution. A gas was

only evolved when specific conditions were met; firstly,

the sample had to have been saturated with carbon dioxide.

Secondly, the solvent had to be sodium bicarbonate.

A different aqueous solvent, sodium sulfate, was used as

a negative control for gas evolution. As expected, gas was

evolved in a carbon dioxide saturated solution of 100mM

hydroquinone in 0.5M NaHCO 3

. Additionally, it was

found that a 100mM sesamol solution in 0.5 M NaHCO 3

also saw gas evolution at the same potential. When each of

these redox mediators were in a solution of 0.5M Na 2

SO 4

,

no gas evolution was seen.

3.3 – Fuel Cell Testing

A fuel cell was tested with no redox mediator, a quinone

redox mediator, and a sesamol redox mediator, and relative

carbon dioxide concentrations were measured with FTIR.

When the membrane placed in the fuel cell was soaked

in a solution of 0.5 M NaHCO 3

, little to no difference was

seen in the relative concentrations of carbon dioxide on

the permeate side of the cell even when voltage was applied

and subsequently turned off. The carbon dioxide

peak was identified to be between 2300 and 2330 cm -1 .

The data were baseline shifted to begin at the same transmittance

value because the FTIR reported different initial

transmittance values with each measurement.

When the membrane was soaked in 5mM 2,6-DMHQ+

5mM 2,6-DMBQ, the carbon dioxide peak increased in

size as the voltage was applied to the cell for progressively

longer amounts of time. The most carbon dioxide was

present on the permeate side when a voltage of 2.5V had

been applied for 10 minutes, as represented by the yellow

peak in Figure 8.

Figure 8. FTIR spectra of permeate side of a fuel cell

with membrane soaked in 5mM 2,6-DMHQ+ 5mM 2,6-

DMBQ in .5 M NaHCO 3

. Spectra were taken with no

voltage applied (blue), with 2.5V applied after 1 min

(orange), after 5 min (gray), and after 10 min (yellow).

When the membrane was soaked in a sesamol solution

that had been oxidized electrochemically to form 5mM

2-HMBQ + 5mM 2-HMHQ, a similar carbon dioxide increase

was seen over time (Fig. 9). The largest peak was

seen when the fuel cell had been running for ten minutes.

Additionally, when the voltage was turned off for 5 minutes,

the concentration of carbon dioxide decreased again,

above any peak that occured when voltage had been applied.

Figure 7. Normalized FTIR spectra of permeate side

of a fuel cell with membrane soaked in 0.5 M NaH-

CO 3

(no redox mediator). Spectra were taken with no

voltage applied (blue), with 2.5 V applied after 1 min

(orange), after 5 min (gray), after 10 min (yellow) and

after the voltage had been turned off for 5 min (light

blue).

Figure 9. FTIR spectra of permeate side of a fuel cell

with membrane soaked in 5mM 2-HMBQ + 5mM

2-HMHQ (generated from the oxidation of sesamol)

in .5 M NaHCO 3

. Spectra were taken with no voltage

applied (orange), with 2.5 V applied after 1 min (gray),

after 5 min (yellow), after 10 min (blue) and after the

voltage had been turned off for 5 min (green).

CHEMISTRY

Broad Street Scientific | 2019-2020 | 37


The sesamol fuel cell saw the greatest increase in peak

height over the ten minutes voltage was applied (Fig. 10).

Although this does not directly correlate to an increase in

fuel cell function, since FTIR testing cannot quantify efficiency,

it does indicate that sesamol is a functional redox

mediator that is qualitatively on par with quinone.

Figure 10. Graph representing peak height vs. time

the voltage has been applied measured from the initial

peak height. Compiled data from bicarbonate

fuel cell (green), quinone fuel cell (blue), and sesamol

fuel cell (yellow).

4. Discussion

The cyclic voltammograms of sesamol above (Fig. 4

and 5) support the proposed scheme for the oxidation of

sesamol, which involves the creation of a quinone species.

The presence of shoulder peaks points to the creation of

a new compound (2-hydroxymethoxybenzoquinone) in

the first step of the reaction, and the subsequent reversible

reaction these compound(s) undergo. When the sample

is purged with argon, (Fig. 4b and 5b), none of the voltammograms

change shape dramatically, suggesting that

none of the peaks were due to the presence of oxygen in

the sample. Additionally, carbon dioxide does not interfere

with the progression of the reaction, as when the samples

are purged with carbon dioxide, the voltammograms

retain their general shape (Fig. 4c and 5c). Although the

cyclic voltammograms in bicarbonate have less defined

peaks (see Fig. 3), this may be attributed in part to the increased

pH and in part to the drift that Ag/AgCl reference

electrodes undergo as they age. Further research could be

done into the impact of pH on sesamol’s redox reaction to

determine if the movement of the peaks represents the impact

of an increased concentration of OH- ions available,

as these are required for the first irreversible oxidation

step of the reaction, and decreased concentration of H+

ions available, as these are required for the reduction of

2-HMBQ to 2-HMHQ. By examining how the pH impacts

the progression of the redox reaction, the optimal pH of a

fuel cell could be determined for maximum carbon dioxide

transport [3].

Half-cell testing results fully support the hypothesis that

sesamol undergoes a proton-coupled electron transfer reaction

because in sodium bicarbonate, a carbon dioxide saturated

solution saw gas evolution at a 0.5V potential. This

gas evolution would not be possible without the transfer

of protons creating an acidic pH at the anode, subsequently

driving the release of carbon dioxide dissolved in solution

as a gas, observed on the electrode. It is important to note

that since only a 0.5V potential was applied, none of the

gas evolution would be expected to be due to water-splitting

[7]. Additionally, the platinum catalyst was chosen

because it does not contribute heavily to water-splitting,

even at higher potentials [3].

Fuel cell testing suggests that quinone and sesamol redox

mediators, rather than the simple application of voltage,

are responsible for moving carbon dioxide across the

fuel cell. Additionally, the decrease in concentration observed

when the voltage is removed dispels the theory that

carbon dioxide is leaking over to the permeate side of the

cell over time. Further work is necessary to evaluate the

efficiency of each fuel cell, as this project’s set-up was not

equipped to evaluate percent carbon dioxide transported,

only observe relative decreases or increases in concentration.

Plotting peak height against time reveals that sesamol

and quinone have similar efficiencies in this setting. All evidence

of carbon dioxide transport was qualitative. Percent

carbon dioxide efficiency can be evaluated with different

applied potentials as well, not just 2.5V. Further research

can also be done to verify that the mechanism is selective

for carbon dioxide by pumping a mixture of nitrogen, oxygen

and carbon dioxide, rather than pure carbon dioxide,

across a fuel cell to further emulate flue gas conditions.

One area of concern is that the platinum catalyst will catalyze

water-splitting and that oxygen could be released on

the permeate side with carbon dioxide, although previous

works have found that platinum on carbon black does not

generate oxygen in large amounts as other similar metal

catalysts might [3].

5. Conclusion and Future Work

It has been shown that sesamol undergoes a quasi-reversible

redox reaction in sodium bicarbonate; it is assumed

that two non-toxic quinones are created, 2-hydroxymethoxybenzoquinone

and 2-hyroxymethoxyhydroquinone,

that further reduce and oxidize reversibly while transfering

protons. The appearance of shoulder peaks in the cyclic

voltammograms supports this reaction scheme, and their

relative size and placement indicate that such a reaction is

reversible and repeatable. Through half-cell testing, it was

confirmed that protons are transferred, thus creating an in

situ pH gradient that releases carbon dioxide that has been

dissolved in solution at the anode. Based on the results of

preliminary fuel cell testing, it can be concluded that sesamol’s

quasi-reversible proton-coupled electron transfer

reaction functions to transport carbon dioxide across a fuel

cell like that of a quinone, and can be used as a more environmentally-friendly

choice of mediator to separate CO 2

38 | 2019-2020 | Broad Street Scientific CHEMISTRY


from flue gas.

6. Acknowledgments

The author would like to thank the North Carolina

School of Science and Mathematics (NCSSM) Science Department,

the NCSSM Foundation, the NCSSM Summer

Research and Innovation Program and research mentor

Dr. Michael Bruno. Additionally, this research would not

have been possible without the generous gift of polypropylene

membrane from Celgard (Charlotte, NC).

7. References

[1] Wuebbles, D. (2017). USGCRP: Climate Science Special

Report: Fourth National Climate Assessment, Volume

I. U.S. Global Change Research Program. DOI: 10.7930/

J0J964J6.

[2] Rheinhardt, J. (2017). Electrochemical Capture and

Release of Carbon Dioxide. ACS Energy Letters 2 (2), 454-

461 DOI: 10.1021/acsenergylett.6b00608

[3] Watkins, D. (2015). Redox-Mediated Separation of

Carbon Dioxide from Flue Gas. Energy & Fuels 29 (11),

7508-7515 DOI: 10.1021/acs.energyfuels.5b01807

[4] Gandhi, M. (2018). In Situ Immobilized Sesamol-Quinone/Carbon

Nanoblack-Based Electrochemical Redox

Platform for Efficient Bioelectrocatalytic and Immunosensor

Applications. ACS Omega 3 (9), 10823-10835 DOI:

10.1021/acsomega.8b01296

[5] Bolton, J (2017). Formation and Biological Targets of

Quinones: Cytotoxic versus Cytoprotective Effects. Chemical

Research in Toxicology 30 (1), 13-37, DOI: 10.1021/

acs.chemrestox.6b00256

[6] Brito, R. (2014). Elucidation of the Electrochemical

Oxidation Mechanism of the Antioxidant Sesamol on a

Glassy Carbon Electrode. Journal of the Electrochemical

Society 161 (5), 27-32, DOI: 10.1149/2.028405jes

[7] Stretton, T. (n.d.). Standard Reduction Table. Retrieved

November 13, 2019, from http://www2.ucdsb.on.ca/tiss/

stretton/Database/Standard_Reduction_Potentials.htm.

CHEMISTRY

Broad Street Scientific | 2019-2020 | 39


MODELING THE EFFECT OF CHEMICALLY MODIFIED

NON-ANTIBIOTIC TETRACYCLINES WITH β-AMYLOID

FIBRILS TO TREAT ALZHEIMER’S DISEASE

Rachel Qu

Abstract

Alzheimer’s Disease is a neurodegenerative disorder in which memory and comprehensive abilities are lost over time.

There is currently no known cure, but the disease has been linked to the aggregation of extracellular β-amyloid plaques.

The tetracycline family of antibiotics has been shown to reduce plaque formation, but use in more complex treatments

involves the risk of bacterial resistance. This project explores the use of tetracycline’s non-antibiotic analogs to reduce

β-amyloid aggregation. Certain chemically modified non-antibiotic tetracyclines (CMTs) were selected to be modeled

alongside the known β-amyloid aggregation inhibitors tetracycline, doxycycline, and minocycline. These were then analyzed

computationally using Molegro to predict the binding affinities of certain CMTs to the β-amyloid protein fibril.

CMT-3 (6-deoxy-6-demethyl-4-dedimethylamino tetracycline), CMT-4 (7-chloro-4-dedimethylamino tetracycline),

CMT-5 (tetracycline pyrazole), and CMT-7 (12-deoxy-4-dedimethylamino tetracycline) were seen to bind more effectively

than known inhibitors. The same compounds were then analyzed using StarDrop, helping to determine how effective

the compounds could perform as oral drugs, and CMT-3 and CMT-7 were suggested to be more suitable in acting as

oral drugs. This information was then used in studies with transgenic Caenorhabditis elegans to confirm results. Treatment

in more complex models, like vertebrates, could be applied in the future to develop a novel treatment method for Alzheimer’s

Disease.

1. Introduction

1.1 – Alzheimer's Disease

Alzheimer’s Disease (AD), the most common type of

dementia, is a neurodegenerative disorder in which memory

and comprehensive abilities are slowly lost over time

[1]. Most symptoms appear in more elderly individuals,

with symptoms generally appearing after age 60 [2]. Features

of the disease include the loss of connections between

neurons. Damage appears to initially take place in the hippocampus,

spreading outward as progression occurs [3].

By 2060, the number of Americans with the disease is projected

to hit around 14 million. Currently, effective prevention

methods do not exist, as there is no known cure

for Alzheimer’s [2]. This is detrimental because AD is one

of the highest ranking causes of death in the United States.

AD occurs when brain cells that typically process, store,

and retrieve information degenerate and die [4]. There are

two supposed causes for this disease, traced back to β-amyloid

(Aβ) peptides and tau proteins. The accumulation of

intracellular neurofibrillary tangles (NFTs), composed of

tau proteins, used to stabilize microtubules when phosphorylated,

is associated with AD. In addition, extracellular

plaques are also exhibited in patients with AD. Aβ

peptides, created from the breakdown of amyloid precursor

protein (APP) primarily form these plaques [5]. The

amyloid hypothesis assumes that mistakes in the process

governing the production, accumulation, and/or disposal

of Aβ proteins are the primary cause of AD. These proteins

accumulate in the brain, disrupting communication

between brain cells and killing them.

1.2 – β-Amyloid Fibrils

Amyloid precursor protein (APP) is cut by other proteins

into smaller separate sections. One of these sections

becomes Aβ proteins, which tend to accumulate. It is currently

believed that small, soluble aggregates of Aβ are

more toxic [4]. First, they form small clusters, or oligomers,

and then chains of clusters called fibrils. The fibrils

then form mats called β-sheets. Finally, the mats come

together to form the plaques seen in AD [4]. Because the

cleavage process by γ-secretase that forms Aβ is not always

entirely precise, Aβ peptides of different lengths can be

formed [6]. Aβ-42 is one of these, thought to be especially

toxic [3].

Because Aβ production and its subsequent fibril formation

is assumed to be a cause of AD, inhibiting Aβ aggregation

could lead to a potential cure for AD. Since Aβ’s structure

relies on hydrogen bonds for β-sheet mat formation,

disruption by certain compounds can potentially be used

to prevent aggregation. The alternating hydrogen bond

donor-acceptor pattern has been thought to be complementary

to the β-sheet conformation of Aβ-42 [7]. Previous

compounds known to act as alleviants or deterrents

of AD have been seen to contain certain common features

that suggest the hydrogen bonds donor-acceptor-donor

patterns are responsible for interrupting the β-sheet formation

of Aβs [7].

40 | 2019-2020 | Broad Street Scientific CHEMISTRY


1.3 – Tetracycline and CMTs

Even though there is no definitive cure, there are still

treatment options that exist to alleviate AD. Many are

natural products with medicinal properties, but other options

also exist. Some compounds like aged garlic extract,

curcumin, melatonin, resveratrol, Ginkgo biloba extract,

green tea, and vitamins C and E have been tested in Alzheimer’s

patients and show promise [8-9]. Tetracycline and

its analogs like doxycycline and minocycline have also been

shown to do the same [10].

Observing tetracycline and some of its analogs showed

similarities in structure between the three compounds.

These specifically included a pattern of alternating hydrogen

bond donors and acceptors attached to adjacent rings

on all three compounds. It was hypothesized that this alternating

donor-acceptor-donor pattern had some basis in

affecting Aβ fibrils, since tetracyclines have been shown

to dissolve these structures after formation [10]. This was

based on the fact that hydrogen bond interactions hold

the β-sheet’s pleated sheet structure together. Since the

donor-acceptor-donor pattern mimics the patterns in the

sheet, it is possible that structures like tetracyclines can disrupt

this hydrogen bonding and thus the structure of the

sheet itself. By disrupting hydrogen bonds in β-sheets, the

protein no longer retains its shape and theoretically loses

its original structure. By this basis, the fibril formation

would be disrupted or reversed and Aβ would not be able

to continue to form its characteristic plaques.

Though tetracycline, doxycycline, and minocycline

have been shown to be effective, they all exhibit antibacterial

characteristics [10]. Antibiotic activity is linked to the

presence of the dimethylamine group present on the structure

(Fig. 1). This could lead to poor side effects, specifically

bacterial antibiotic resistance, if clinically used to treat

AD. Removal of this group while maintaining the primary

structure of the donor-acceptor-donor pattern would ideally

lead to the formation of an improved AD drug.

Chemically modified non-antibiotic tetracyclines

(CMTs) have been synthesized and can be applied to Aβ

aggregation inhibition. They have similar structures to

tetracycline but remove the dimethylamine group responsible

for antibiotic properties. Since studies have shown

that tetracycline and its analogs can be effective through

preventing and inhibiting Aβ aggregation, the possible use

of CMTs that still contain the alternating H-bond donor/

acceptor pattern but do not have antibiotic properties is of

interest.

CHEMISTRY

1.4 – Hypothesis/Goals

It is hypothesized that the treatment of AD can be approached

through the use of CMTs to reduce Aβ aggregation,

which can be modeled both computationally and in

vivo through the use of Caenorhabditis elegans. The goals of

this project were to first computationally model Aβ interactions

with different compounds using Molegro to evaluate

aggregation inhibition efficiency, and then evaluate the

same compounds for effectiveness as oral drugs in Star-

Drop. After computational modeling, selected compounds

were to be used with a transgenic version of the model organism

C. elegans to test the in vivo effects of selected compounds

through a paralysis assay.

Figure 1. Structure of tetracycline indicating alternating

hydrogen bond donor-acceptor-donor

pattern (circled in red) and dimethylamine group

(squared in blue).

2. Materials and Methods

Molegro Virtual Docker and Molegro Data Modeller

were used to analyze the binding affinity of multiple chemically

modified tetracyclines (CMT-1, CMT-3, CMT-4,

CMT-5, CMT-6, CMT-7, CMT-8) to Aβ proteins. Aβ

structure 2MXU was taken from the Protein Data Bank.

CMT structures were created in ChemDraw and translated

to PDB files by putting the SMILES code through an

online translator. After docking CMT structures, their

Ligand Rerank scores were converted to binding affinity

scores using Molegro’s Binding Affinity algorithm. Star-

Drop was used to analyze the ability of the drug to reach

the target cells using Lipinski’s Rule of Five and an Oral

CNS scoring profile. Data from Molegro and StarDrop

were taken into account to rank the drug performance and

narrow the list of potential drugs.

CL2006 strain of Caenorhabditis elegans was obtained

from the Caenorhabditis Genetics Center at the University

of Minnesota and grown on Nematode Growth Medium

plates. CMT-3 was purchased from Echelon BioSciences.

All other chemicals were obtained from Sigma Aldrich and

dissolved in water before use.

C. elegans were propagated at 15 degrees Celsius on solid

Nematode Growth Medium (NGM). Age synchronization

was performed by picking worms of the same life stage,

L1s, and plated on plates seeded with tet-resistant zim-2

Escheria coli. After 48 hours, plates were treated with 100

μL of vehicle or drugs (100 μM each). After 24 hours, the

temperature was raised to 24 degrees Celsius and paralysis

was scored until 120 hours of age. Paralysis was evaluated

Broad Street Scientific | 2019-2020 | 41


by touching the worms’ heads; paralysis was scored if the

worm did not move or only moved the head.

3. Results/Discussion

3.1 – Molecular Docking Studies

Molegro is a computational platform for predicting

protein-ligand interactions. In Molegro, the protein

2MXU was taken from the Protein Data Bank as a model

of Aβ fibril aggregation. 2MXU shows a fibril structure

composed of multiple peptide monomers. Two different

approaches were taken with this protein. One was of a

smaller fibril structure and the other of a dimer with a cavity

located between. Because the mechanism behind how

drugs like tetracycline work to dissolve the preformed fibril

is not currently known, both approaches were used.

With the fibril structure approach, the ligand bonds to the

outside of the fibril protein in a smaller cavity. With the

dimer structure approach, the ligand binds between the

two monomers (Fig. 2).

binding when more negative. Through using Molegro to

model the interactions between protein and ligand, it was

possible to find that in both the fibril and dimer formation,

CMTs repeatedly and consistently outperform tetracycline

and its natural analogs (Tab. 1).

Table 1. Molegro data table of binding affinities,

where more negative values are more effective, for

tetracycline and analogs with both the fibril (top)

and dimer (bottom) approaches.

Molecule Fibril binding affinity (avg.)

Tetracycline -9.69

Doxycycline -10.10

Minocycline -8.28

CMT-3 -14.43

CMT-1 -13.45

Molecule Dimer binding affinity (avg.)

Tetracycline -11.00

Doxycycline -12.33

Minocycline -9.37

CMT-3 -17.17

Among the CMTs themselves, however, there were

varied results. By testing different CMTs, it was possible

to find the more effective Aβ aggregation inhibitor to use

in further treatment. CMTs-1, 3, 4, 5, 6, 7, and 8 were all

used (Fig. 3). Because the same trend was seen in both the

fibril and dimer approach, with CMTs outperforming tetracycline

and its natural analogs, the calculated binding

affinities were only compared for the fibril approach. The

compounds with the most effective binding affinities were

shown to be CMT-3, CMT-4, CMT-5, and CMT-7 (Tab.

2).

Figure 2. Molegro screencap of Aβ fibril protein

2MXU with tetracycline (green) integrated into both

fibril structures (top) and dimer structure (bottom).

When modeling the effectiveness of different ligands on

Aβ aggregation inhibition efficiency, tetracycline and its

analogs can be interpreted as effective through the analysis

of binding affinities. Binding affinities show more effective

Figure 3. Structures of CMTs (in order: 1, 3, 4, 5, 6, 7, 8).

42 | 2019-2020 | Broad Street Scientific CHEMISTRY


Table 2: Molegro data table of binding affinities for

CMT's using the fibril approach

Molecule

CMT-1 (4-dedimethylaminotetracycline)

CMT-3 (6-deoxy-6-demethyl-4-dedimethylaminotetracycline)

CMT-4 (7-chloro-4-de-dimethylamino

tetracycline)

CHEMISTRY

Binding affinity

(average)

-13.45

-14.43

-17.03

CMT-5 (tetracycline pyrazole) -16.01

CMT-6 (4-dedimethylamino.

4-hydroxytetracycline)

CMT-7 (12-deoxy-4-de-dimethylamino

tetracycline)

CMT-8 (4-dedimethylaminodoxycycline)

-11.32

-15.26

-14.09

3.2 – Drug Efficiency Analysis

StarDrop was used in determining how the different

CMTs would perform as oral drugs in treating AD. Lipinski’s

Rule of Five and an Oral Central Nervous System

(CNS) Profile were both performed. Lipinski’s Rule of

Five evaluates if a compound will be a likely active drug

in humans. The criteria include a maximum value of 5 for

octanol-water partition coefficient (logP) which measures

hydrophobicity, a maximum of 500 Da molecular weight

(MW), a maximum of 5 hydrogen bond donors (HBD)

and 10 hydrogen bond acceptors (HBA). It was shown

that CMT-3 and CMT-7 were the only two CMTs that

fulfilled all four criteria, with even tetracycline and doxycycline

only meeting three of the four requirements (Tab.

3). Since tetracycline, doxycycline, and minocycline are all

used in practice as drugs, Lipinski’s Rule of Five is only one

guideline for drug design and not absolutely required for

effective drugs. However, it can still be used as a guideline

to corroborate the potential effectiveness of CMT-3 and

CMT-7.

An Oral CNS profile was also performed using Star-

Drop. Because AD is related to the brain, the drugs need

to be able to be analyzed in the context of the central nervous

system. While none of the compounds exhibited ideal

Oral CNS profiles, many outperformed tetracycline. Since

tetracycline is known to be able to be used in the context

of treating the brain [10], the criteria used for a drug to

be deemed effective was an Oral CNS Profile score higher

than that of tetracycline. This allowed CMT-3, 4, 5, 7, and

8 to all be predicted as effective Oral CNS drugs (Tab. 4).

From the results of StarDrop and Molegro combined,

it was determined that CMT-3 and CMT-7 were the best

candidates for binding to the fibril and for being an effective

drug capable of treating AD through the context of Aβ

aggregation. This information was then applied to experimental

treatments.

Table 3: StarDrop data table of CMTs, tetracycline,

doxycycline, and minocycline through the analysis

of Lipinski’s Rule of Five. The color scale ranges

from green to red, with darker greens being indicators

of more effective drugs, and darker reds being

indicators of less effective drugs.

Lipinski

Rule of

Five

Compound logP MW HBD HBA

0.5 tetracycline -0.8803 444.4 6 10

0.5 doxycycline -0.2456 444.4 6 10

1 minocycline 0.4239 457.5 5 10

0.5 CMT-1 -0.3233 401.4 6 9

1 CMT-3 0.7645 371.3 5 8

0.5 CMT-4 0.9379 435.8 6 9

0.5 CMT-5 -0.3333 397.4 6 9

0.5 CMT-6 -0.9992 417.4 7 10

1 CMT-7 0.4774 385.4 5 8

0.5 CMT-8 0.1577 401.4 6 9

3.3 – Caenorhabditis elegans

Caenorhabditis elegans (C. elegans) do not naturally express

Aβ. However, transgenic strains of the nematode

exist in which it is possible to model the effectiveness of

Alzheimer’s drugs, using temperature-sensitive expression

strains. The CL2006 strain of C. elegans, when raised

at normal conditions, expresses typical wild-type growth.

When shifted to higher temperatures, they conditionally

express the Aβ peptide and upon adulthood, paralysis

is induced. By treating C. elegans with AD drugs, the percentage

of nematodes in which paralysis is expressed will

decrease, as Aβ plaques are lessened. The more effective

the drug is, the more paralysis should be reduced. Previous

studies with tetracycline show that, 48 hours after the

temperature shift, around 35 percent of C. elegans treated

with doxycycline, minocycline, and tetracycline remained

healthy while 100 percent of untreated worms were paralyzed

[10].

A similar paralysis assay for the CL2006 strain of C. elegans

was performed with CMT-3 and it was found that

CMT-3 was more effective at preventing paralysis than

tetracycline and doxycycline, with doxycycline being the

least effective (Fig. 4).

3.4 – Discussion

Molegro and StarDrop data indicated that the most effective

CMTs to treat AD would be CMT-3 and CMT-7,

properly known as 6-deoxy-6-demethyl-4-dedimethyl-

Broad Street Scientific | 2019-2020 | 43


Table 4. StarDrop data table of CMTs, tetracycline, doxycycline, and minocycline through the analysis of an

Oral CNS Scoring Profile. The color scale ranges from green to red, with darker greens being indicators of

more effective drugs, and darker reds being indicators of less effective drugs.

Oral

CNS

Scoring

Profile

Compound logP logS 2C9

pKi

hERG

pIC50

BBB

log([brain])

:([blood])

BBB

category

HIA

category

P-gp

category

2D6

affinity

category

PPB90

category

0.03053 tetracycline -0.8803 3.002 4.742 2.599 -1.577 - - yes low low

0.08814 doxycycline -0.2456 2.833 4.739 2.659 -1.3 - + yes low low

0.1369 minocycline 0.4239 2.591 4.8 3.168 -1.221 - + yes low low

0.02148 CMT-1 -0.3233 1.843 4.91 2.232 -1.217 - - yes low low

0.03198 CMT-3 0.765 1.369 4.833 2.665 -1.094 - - yes lo low

0.06815 CMT-4 0.9379 1.837 5.02 2.371 -1.198 - + yes low low

0.03579 CMT-5 -0.3333 2.412 5.232 2.096 -1.552 - - yes low low

0.01162 CMT-6 -0.9992 2.177 4.971 2.035 -1.3 - - yes low low

0.03813 CMT-7 0.4774 1.556 4.886 2.429 -0.9871 - - yes low low

0.07435 CMT-8 0.1577 1.73 4.88 2.422 -1.206 - + yes low low

amino tetracycline and 12-deoxy-dedimethylamino tetracycline

respectively. CMT-3 is commercially available as

incyclinide while CMT-7 is not, so CMT-3 was the only

CMT taken into consideration in experimental data. In future

work, if other CMTs are available, they should also

be used to measure effectiveness as AD drugs, especially

CMT-7.

Figure 4. Paralysis assay of C. elegans paralysis 24

hours after temperature shift to 24 o C.

It has been shown that tetracycline and doxycycline are

effective at reducing paralysis in C. elegans [10]. Results

shown here support this while suggesting that CMT-3 is

also effective at reducing paralysis, potentially even better

than the two tetracycline analogs.

However, though data suggest that CMT-3 is overall

more effective than tetracycline and doxycycline, the data

show large drops and fluctuations during the time period

measured, probably due to a smaller sample size in each

trial performed. Additionally, there is a 10-hour long time

period lost in data collection that distorts the pattern of

paralysis as time goes on. In future work, the experimental

section should be replicated with a larger sample size and

ideally consistent 2-hour intervals throughout the entire

approximately 24-hour period, working to confirm the results

shown here.

4. Conclusion

It has been concluded through computational analysis

that CMTs can be predicted to interact with Aβ in a similar

manner to tetracycline and its natural analogs. Especially

successful CMTs include CMT-3 and 7, and in vivo

trials with model organism C. elegans, specifically temperature-dependent

strain CL2006, can model paralysis correlated

to Aβ expression. Experimental data shows that

CMT-3 does partially recover paralysis in this strain and

suggests that other CMTs may do the same.

Tetracycline and its analogs can be used for multiple

medicinal applications and can interact with Aβ aggregation

in transgenic models of C. elegans to protect them

from the paralysis phenotype. Certain CMTs have been

computationally modeled to behave in the same manner

to a similar and possibly better degree. This is especially

applicable to CMT-3 (6-deoxy-6-demethyl-4-dedimethylamino

tetracycline) and CMT-7 (12-deoxy-4-dedimethylamino

tetracycline) which have been suggested to be

better at preventing aggregation with the preformed Aβ

fibril and exhibit better interactions as oral drugs. Paralysis

assays performed with a transgenic model of C. elegans in

an in vivo study seem to exhibit a pattern indicating CMT-

3’s ability to prevent paralysis caused by Aβ aggregation

and may outperform both tetracycline and doxycycline in

44 | 2019-2020 | Broad Street Scientific CHEMISTRY


its role as an aggregation inhibitor, though future replications

must be done to corroborate this phenomenon.

CMT-3 is therefore a promising compound to be used

in future studies as a proposed treatment method for preventing

or curing AD. It would be done through the same

method tetracycline and its natural analogs exhibit, while

still preventing associated bacterial antibiotic resistance,

and reducing side effects that would occur in patients

through treatment of AD with tetracycline.

5. Acknowledgements

This research was supported by the North Carolina

School of Science and Mathematics Science Department,

the NCSSM Foundation, and the NCSSM Summer Research

and Innovation Program. The author would also

like to thank research mentor Dr. Michael Bruno and Dr.

Kim Monahan for her assistance in C. elegans research.

6. References

[1] Xu, J., Kochanek, K. D., Sherry, M. A. ;, Murphy, L.,

& Tejada-Vera, B. (2007). National Vital Statistics Reports,

Volume 58, Number 19 (05/20/2010). Retrieved from http://

www.cdc.gov/nchs/deaths.htm.

Biological Chemistry, 287(41), 34786–34800. https://doi.

org/10.1074/jbc.M112.357665

[8] Bui, T. T., & Nguyen, T. H. (2017, September 26).

Natural product for the treatment of Alzheimer’s disease.

Journal of Basic and Clinical Physiology and Pharmacology,

Vol. 28, pp. 413–423. https://doi.org/10.1515/

jbcpp-2016-0147

[9] Wu, Y., Wu, Z., Butko, P., et al. (2006). Amyloid-β-induced

pathological behaviors are suppressed

by Ginkgo biloba extract EGB 761 and ginkgolides in

transgenic Caenorhabditis elegans. Journal of Neuroscience,

26(50), 13102–13113. https://doi.org/10.1523/JNEURO-

SCI.3448-06.2006

[10] Diomede, L., Cassata, G., et al. (2010). Tetracycline

and its analogues protect Caenorhabditis elegans from β

amyloid-induced toxicity by targeting oligomers. Neurobiology

of Disease, 40(2), 424–431. https://doi.org/10.1016/j.

nbd.2010.07.002

[2] Matthews, K. A., Xu, W., Gaglioti, A. H., Holt, J. B.,

Croft, J. B., Mack, D., & McGuire, L. C. (2019). Racial

and ethnic estimates of Alzheimer’s disease and related

dementias in the United States (2015–2060) in adults aged

≥65 years. Alzheimer’s and Dementia, 15(1), 17–24. https://

doi.org/10.1016/j.jalz.2018.06.3063

[3] National Institute on Aging (2017). What Happens

to the Brain in Alzheimer's Disease? National Institute

on Aging. https://www.nia.nih.gov/health/what-happens-brain-alzheimers-disease

[4] Alzheimer's Association (2017). Beta-amyloid and the

amyloid hypothesis. Alzheimer’s Association. https://www.

alz.org/national/documents/topicsheet_betaamyloid.pdf

[5] Markaki, M., & Tavernarakis, N. (2010). Modeling

human diseases in Caenorhabditis elegans. Biotechnology

Journal, 5, 1261–1276. https://doi.org/10.1002/

biot.201000183

[6] Murphy, M. P., & Levine, H. (2010). Alzheimer’s

Disease and the Amyloid-β Peptide. Journal of Alzheimer’s

Disease, 19(1), 311–323. https://doi.org/10.3233/jad-

2010-1221

[7] Kroth, H., Ansaloni, A., et al. (2012). Discovery and

structure activity relationship of small molecule inhibitors

of toxic β-amyloid-42 fibril formation. Journal of

CHEMISTRY

Broad Street Scientific | 2019-2020 | 45


SYNTHESIS OF A TAU AGGREGATION INHIBITOR

RELATED TO ALZHEIMER’S DISEASE

Emma G. Steude

Abstract

A multitargeted approach is suggested to be most effective in inhibiting the formation of tau aggregates in Alzheimer’s

disease. Two promising targets for treatment of Alzheimer’s are the initial hyperphosphorylation of tau, caused by an

overexpression of the GSK-3β protein, and early tau aggregation itself. To improve drug effectiveness, the structure of a

known inhibitor molecule targeting both of these stages of tau aggregation was adjusted to increase binding affinity with

the GSK-3β enzyme. These adjusted molecules were screened using Molegro. The candidate molecules with the highest

calculated binding affinities were further evaluated. One of these novel compounds was then synthesized and assayed for

its ability to inhibit the GSK-3β protein, resulting in a comparable efficacy to the original known molecule’s multitargeted

structure. The novel molecule has promising GSK-3β inhibition results and maintained structural features to attack

early tau aggregation. This indicates possible effectiveness in inhibiting the future stages of tau aggregation indicative of

Alzheimer’s disease.

1. Introduction

According to the Alzheimer’s Association [1], one in

ten Americans aged 65 and older have Alzheimer’s disease.

The neural damage from Alzheimer’s disease can result in

severe memory loss, degradation of motor functions, and

eventual death. Despite the severity of the disease, there is

currently no cure. This is partly because a specific target

has yet to be definitively identified and experimentally determined

to cause Alzheimer’s disease. Currently, the two

most researched explanations are the amyloid hypothesis

and the tau hypothesis. The amyloid hypothesis suggests

that amyloid-β aggregates to create plaques that disrupt

normal neuron communication at the synapses. However,

none of the developed amyloid beta-targeted drugs have

proven to stop or even slow disease progression [2]. More

recently, the tau hypothesis is being researched. Tau is

thought to disrupt transport in the neuron itself and may

be a more promising target for prevention of Alzheimer’s

disease.

The tau protein is located along the axon of neural cells

and is complex, having six isoforms. In healthy brains, tau

functions to aid transport of signals across the axon of

neurons in the brain. In the brains of Alzheimer’s disease

patients, however, tau hyperphosphorylates, causing the

tau to break from the microtubule that it was stabilizing.

Drifting from its usual position near the axon, the tau protein

can then form aggregates with other hyperphosphorylated

tau proteins [3]. The formation of tau aggregates is

thought to be a significant factor in causing Alzheimer’s

disease.

There are, in fact, multiple levels of tau aggregation

and, consequently, multiple targets for drug design. Hyperphosphorylated

tau proteins may aggregate to form

β-sheets, which later aggregate to form oligomers. Soluble

oligomers then form insoluble paired helical filaments

(PHF), which further aggregate to form neurofibrillary

tangles [3]. Researchers are still unsure which specific aggregation

level would cause Alzheimer’s disease, but experimental

correlation suggests that overall aggregation

plays a significant role in the progression of the disease.

As there is no direct target for disease treatment, focusing

on early stages of the aggregation process may be the

best alternative. The inhibition of early aggregation may

prevent subsequent, more complex aggregates from forming.

Furthermore, targeting multiple steps of aggregation

with a multitargeted drug may be more effective in preventing

the disease than targeting a single stage. The hyperphosphorylation

stage and the early levels of aggregation

itself are two plausible targets. Drug interaction with

these targets can be tested with the GSK-3β protein and

the AcPHF6 peptide. GSK-3β is a kinase that, when overexpressed,

hyperphosphorylates the tau protein, enabling

tau to aggregate. The AcPHF6 peptide, on the other hand,

is a segment of the tau protein that models aggregation.

This peptide is involved in both the microtubule-binding

property of normal tau as well as PHF formation in hyperphosphorylated

tau. Therefore, one can inhibit many

stages of tau protein aggregation by inhibiting the GSK-

3β protein’s hyperphosphorylation of tau and by inhibiting

the early aggregation shown through the AcPHF6 peptide.

In essence, by targeting tau aggregation at two early levels,

further aggregation may be effectively inhibited [4].

Previous research used Thiadiazolidinedione (TZD) as

a lead compound to identify candidate compounds that act

against both the GSK-3β protein and the AcPHF6 peptide

[4]. One derived molecule, Model 30 [4] was promising

for multitargeted tau inhibition (Fig. 1). It was suggested

that structural adaptations at the R1 and R2 positions

could further increase the efficacy of this molecule. With

these adaptations, the compound is hypothesized to more

effectively inhibit tau aggregation at the GSK-3β target

46 | 2019-2020 | Broad Street Scientific CHEMISTRY


site, while preserving the characteristics necessary to inhibit

the AcPHF6 peptide. This work aims to confirm this

hypothesis by developing a more potent variant of this

derivative with a greater efficacy against GSK-3β while

maintaining the structural features thought to be required

for inhibition of the AcPHF6 peptide.

Figure 1. Structure of Model 30, with suggested structural

adaptations at R1 and R2.

2. Methods

2.1. Inhibitor Design through Computational Modeling

TZD derivative Model 30 was selected as a starting

point for computational modeling [4]. Candidate molecules

were ultimately designed by selecting modifications

that followed the proposed structural adaptations (Fig. 2).

Figure 3. The active site pocket of the GSK-3β protein,

with identifiers Lys85 and Val135.

To assess the candidates’ suitability in medical application,

the molecules with the best predicted effectiveness

were then imported into StarDrop. The Oral CNS Scoring

Profile was run in StarDrop and the results were reviewed,

taking note of the Lipinski Rule of Five data as well as

the blood brain barrier (BBB) log([brain]:[blood]) value.

The StarDrop data for the newly designed structures were

compared to those of the original compound, Model 30.

The candidates that compared most favorably to Model 30

were selected for possible synthesis.

2.2 – Synthesis, Purification, and Analysis

The reaction proceeds by reacting the appropriate aldehyde

precursor with TZD (1:1) under microwave irradiation

at 80 °C for 30 minutes with half the molar amount of

Ethylenediamine diacetate (EDDA) to catalyze the reaction

(Fig. 4) [4]. The chosen compound, T006129, was synthesized

using this method with its respective R-group (Fig.

5).

Figure 2. Candidate molecules designed with base

structure of Model 30.

The candidate structures were then imported into Molegro

Virtual Docker along with the GSK-3β protein structure

[5]. Cavities on the protein were detected, and the active

site was identified to be around Val135 and Lys85 on

the protein (Fig. 3). The candidates were then docked in

the active site of GSK-3β. Poses were compared using the

binding affinity scores, where more negative scores suggested

better binding effectiveness.

Figure 4. Reaction scheme proposed to synthesize the

proposed inhibitors [4].

The thick, caramel-colored product was then diluted

with water and collected by filtration, being sure to scrape

out as much of the mixture as possible. After vacuum filtering

with a heavy wash of water, the compound was then

purified by crystallization. To achieve this, a considerable

amount of hot ethanol was added by pipette to the solution

of product, which was heated and stirred until fully

dissolved. Then, the solution was taken off the hot plate

CHEMISTRY

Broad Street Scientific | 2019-2020 | 47


and chilled water was added by drop until crystals began

to form. Once this occurred, the beaker was carefully set

in ice and not disturbed to allow for crystallization to complete.

The crystals were collected by vacuum filtration

and dried. Thin layer chromatography (TLC) was used to

monitor the progress of the reaction.

in solution (1mM). The assay protocols were adjusted to

fill cuvettes for analysis by fluorescence emission in a spectrophotometer.

Four solutions were prepared and tested using the above

assays. A solution with the GSK-3β enzyme, but no ATP

or inhibitor, was used to create the lower boundary for

expected fluorescence. A solution with ATP, but no GSK-

3β or inhibitor, determined the upper boundary with the

greatest amount of ATP to be expected. Next, a solution

with the enzyme and ATP, but no inhibitor, showed the

regular activity of GSK-3β. Lastly, all three components

were put in the solution to find the impact of the inhibitor

on GSK-3β activity. The % inhibition was calculated for

the specific inhibitor concentration.

3. Results and Discussion

3.1 – Computational Analysis

Molegro Computational Analysis showed a clear increase

in predicted inhibition when the substituents with

the suggested properties were added. The adapted molecules

have better predicted binding affinities compared to

the Model 30 structure (Fig. 6). The more negative scores

suggest a better inhibition for the GSK-3β enzyme.

Figure 5. Respective R-groups for the proposed synthesis

procedure.

Fourier-transform infrared spectroscopy (FTIR) was

then used to determine whether the inhibitor was in fact

created. Heptane, ethyl acetate, and chloroform were unable

to sufficiently dissolve the product for liquid FTIR,

so KBr pellets were created to run FTIR. Since KBr pellets

are susceptible to water vapor, the crushed inhibitor

was mixed with uncrushed KBr before being compressed.

A blank KBr pellet was used as the background to further

reduce the influence of the O-H bond of water on the resulting

graph. FTIR graphs from the reactants and the

product were compared to determine whether the reaction

occurred correctly.

2.3 – Assay to Test GSK-3β Inhibition

The BioVision ATP Colorimetric/ Fluorometric Assay

Kit was paired with the BPS Bioscience GSK-3β Assay Kit

to determine the molecule’s inhibitory effect on the GSK-

3β protein by inducing measurable fluorescence. Properly

functioning GSK-3β consumes ATP in its reaction with a

substrate, so inhibited GSK-3β would leave high ATP levels.

Since the measured fluorescence directly corresponds

to the amount of ATP left in solution, inhibition can be

tested by measuring fluorescence of the solution.

To prepare the product for the assays, just enough dimethyl

sulfoxide (DMSO) was added to dilute the product

Molecule

Inhibitor Binding Affinities

Binding

Affinity

Model 30 -27.886

Molecule

Binding

Affinity

T002052 -30.7393 PyrroleC 3

O 2

-31.7228

T006129 -31.1837 PyrroleNH 2

-29.5741

T010305 -33.3667 T01113w/

OCH 3

-29.3512

Figure 6. Molegro computational binding affinities

for Model 30 and six candidate inhibitors.

With these promising computational inhibition results,

these six candidate compounds were imported into

StarDrop to assess suitability in medical application. The

Lipinski Rule of Five and the Oral CNS Scoring Profile

data were compared between molecules. The compounds

seem to be viable in most of the metrics computed by

StarDrop. See the StarDrop Data in Supplemental Materials,

below. Between molecules, the most variation is

with respect to the blood brain barrier permeability (BBB

log([brain]:[blood])). Compounds with more positive BBB

values can penetrate the BBB more easily, increasing their

efficacy in reaching the target site in the brain and, therefore,

requiring a lower dosage. Disappointingly, when

comparing the BBB permeability to the binding affinity

scores (Fig. 7), there were few compounds that noticeably

stood out from the rest in both areas.

48 | 2019-2020 | Broad Street Scientific CHEMISTRY


the nitrogen on the indole and the carboxylic acid group,

were preserved in the product (Fig. 8). The aldehyde peaks,

on the other hand, were absent in the product, confirming

that the appropriate reaction did occur.

Figure 7. Blood brain barrier permeability compared

to the computational binding affinities for Model 30

and candidate molecules. The upper left data points

are most ideal in each respect.

Generally, candidates with better predicted inhibition

had worse predicted permeability and vice versa. This

made choosing compounds to synthesize more difficult.

Model 30 had the most effective predicted BBB permeability

(Fig. 7), but other molecules had preferable computational

binding affinities. Considering that no compound

stands out for excellent scores in both variables and that a

goal of this particular research project is to more effectively

inhibit GSK-3β particularly, the candidates with more

promising binding affinity scores were selected for possible

synthesis. These were molecules T0103050, PyrroleC 3

0 2

,

T006129, and T002052. Attempting to use the reaction

mechanism previously proposed and taking the availability

of laboratory materials into consideration, candidate molecule

T006129 was then chosen to be synthesized.

3.3 – Assay Analysis

When the GSK-3β inhibition assay was performed,

product T006129 was experimentally determined to affect

the performance of the GSK-3β protein. The molecule had

inhibitory results when added to the reaction solution (Fig.

9). As the experiment was not able to be repeated in the

limited time frame, a true IC50 score for T006129 could

not be determined. Even so, initial results show that inhibitor

T006129 had a 53% inhibition at a 0.98μL concentration,

which is comparable to Model 30’s IC50 score of

0.89 ±0.21μL. The promising GSK-3β inhibition results of

T006129 point to inhibition of hyperphosphorylation of

tau. At the same time, the molecule maintains its structural

features to attack early tau aggregation, suggesting that

T006129 may be effective in inhibiting future stages of tau

aggregation. These results support that this research approach

may yield new and even more effective compounds.

By refining the T006129 inhibitor’s IC50 score with more

GSK-3β assay trials and testing the other promising inhibitor

candidates defined in this study, the GSK-3β inhibition

results may be more accurately compared. This additional

research would pave the way to testing the impact of the

early multitargeted approach on later stages of tau aggregation.

3.2 – Synthesis Analysis

The FTIR analysis of the product shows promising results

for the synthesis of molecule T006129 (Fig. 8). With

a yield of approximately 10%, product T006129 was concluded

to have been produced.

Figure 9. GSK-3β assay fluorescence results for compound

T006129 are shown above. The % inhibition

was found by dividing the amount of ATP used by the

enzyme with the inhibitor present by the amount of

ATP used without the inhibitor present.

4. Conclusion

Figure 8. FTIR data comparing the components of the

T006129 product to the aldehyde reactant.

Prominent absorption lines of the reactant, including

CHEMISTRY

The multitargeted approach to treating Alzheimer’s disease

with the tau hypothesis has yet to be further tested.

While multitargeted molecules have been proposed, focusing

on specialized binding properties may increase drug

effectiveness and decrease undesirable side effects. The

Model 30-derived candidate molecules outlined in this paper

were computationally predicted to have more interaction

with the GSK-3β protein than the baseline Model 30

molecule based on their computed binding affinities. One

Broad Street Scientific | 2019-2020 | 49


such novel molecule was successfully synthesized. The assay

suggests that this inhibitor was comparable to Model

30, with a 53% inhibition at a 0.98μL concentration compared

to a 50% inhibition at a 0.89 ±0.21μL concentration

respectively. Further investigation is required to refine the

measurement of this IC50 score and determine the early

aggregation inhibition as modeled by the AcPHF6 peptide.

A proper IC50 score of inhibitor T006129 could not be

determined in the limited time frame. Even so, the comparable

results suggest that the performance of T006129

and the other inhibitors should continue to be explored.

Among the other candidate compounds explored, compounds

T0103050 and T00252 computationally outperformed

T006129 when taking blood brain barrier permeability

into consideration with binding affinity.

Future studies should refine the IC50 estimate of

T006129 with multiple replicates of assay data. The other

candidate inhibitors should also be synthesized and tested

for their performance against GSK-3β. Additionally, it

would be wise to take factors like the AcPHF6 inhibition

and blood brain barrier permeability into consideration.

These inhibitors are ultimately meant to be multitargeted

drugs. While the focus in this study was to improve the

GSK-3β inhibition specifically, the inhibition of early tau

aggregation and the dosage efficiency of the compound

must be considered in future studies. The results of this

study suggest that a multitargeted inhibitor with improved

GSK-3β inhibition may be created, but a considerable

amount of research is imperative before a treatment for

Alzheimer’s disease can ultimately be proposed.

7. References

[1] Alzheimer’s Association. (2018). 2018 Alzheimer’s

Disease Facts and Figures, Alzheimers & Dementia, 14(3),

367–429. doi: 10.1016/j.jalz.2018.02.001

[2] Myths. (n.d.). Retrieved from http://www.alz.org/

alzheimers-dementia/what-is-alzheimers/myths PDB ID:

1Q4L

[3] Kuret, J., et al. (2005). Pathways of tau fibrillization.

Biochimica Et Biophysica Acta (BBA) - Molecular Basis

of Disease, 1739(2-3), 167–178. doi: 10.1016/j.bbadis.2004.06.016

[4] Gandini, A., et al. (2018). Tau-Centric Multitarget Approach

for Alzheimer’s Disease: Development of First-in-

Class Dual Glycogen Synthase Kinase 3β and Tau-Aggregation

Inhibitors. Journal of Medicinal Chemistry, 61(17),

7640–7656. doi: 10.1021/acs.jmedchem.8b00610

[5] Bertrand, J., et al. (2003). Structural Characterization

of the GSK-3β Active Site Using Selective and Non-selective

ATP-mimetic Inhibitors. Journal of Molecular Biology,

333(2), 393–407. doi: 10.1016/j.jmb.2003.08.031

5. Acknowledgements

I would like to acknowledge Dr. Timothy Anglin, Dr.

Michael Bruno, Dr. Kat Cooper, Dr. Darrell Spells, Mr.

Bob Gotwals, the NCSSM Foundation, the NCSSM Science

Department, the Summer Research and Innovation

Program, Dr. Sarah Shoemaker, and my Research in

Chemistry Peers for making this research possible.

6. Supplementary Materials

6.1 StarDrop Data

Compound T002052 T006129 T010305 Pyrrole

C 3

0 2

Pyrrole

NH 2

T01113w/ Model 30

OCH 3

Oral CNS Scoring

Profile Score

Lipinski Rule of

Five Score

BBB

log([brain]:[blood])

0.1482 0.1183 0.1899 0.1219 0.1478 0.2131 0.2373

1 1 1 1 1 1 1

-0.4836 -1.284 -1.032 -1.284 -0.6675 -0.7091 -0.1947

50 | 2019-2020 | Broad Street Scientific CHEMISTRY


A TELEMETRIC GAIT ANALYSIS INSOLE AND

MOBILE APP TO TRACK POSTOPERATIVE FRACTURE

REHABILITATION

Saathvik A. Boompelli

Abstract

Accurate and objective monitoring of a fracture’s healing process is essential to both patient quality of care and determination

of the chances of nonunion and postoperative intervention. In recent years, due to industrialization, injury rates in

developing countries, notably road traffic injuries (RTIs), have drastically increased. Ill-equipped countries such as Kenya

may have only 60 orthopedic surgeons for a population of 36.9 million, with no rigorous rehabilitation protocol or quality

post-operative care. This work focuses on the development of a telemetric gait analysis insole that works in conjunction

with a mobile application. This technique automates the tedious and resource-intensive process of tracking postoperative

fracture rehabilitation. This is done through analyzing patient ground reaction forces (GRFs), which have been found to

correlate well with weight-bearing ability, fracture healing, and delayed union. Four force-sensitive resistors (FSRs) that

function through polymer thick film technology are placed in the insole under the primary areas for force measurement.

An Arduino microcontroller compiles the data and sends it to a Python program via a Bluetooth module. The Python

program performs peak analysis to determine the average peak Vertical Ground Reaction Force (VGFR) of the strides.

This value is sent to a cloud database that subsequently sends the data to a mobile application made accessible to a healthcare

professional, who can send back weight-bearing recommendations based on the data. Features include access to the

cloud database, a graph of weekly values, emailing functionality, and an ability to begin a trial through a smartphone. This

approach addresses a commonly overlooked lapse in the healthcare system of many countries and provides an objective

method to track fracture healing.

1. Introduction

1.1 – Motivation

1.1.1 – Lower Extremity Injury Prevalence

The economic boom occurring in developing countries

comes at the cost of new and challenging problems, including

the rapid increase of Road Traffic Injuries (RTIs) with

the urbanization of rural areas. There is often a lack of legislation

regulating the education of the local population on

how to navigate the new infrastructure being constructed

in their areas. This results in a higher prevalence of fractures

and lower-extremity injuries in developing countries

(Fig. 1), accounting for more than 1.27 million deaths per

year, and more deaths than HIV/AIDS, tuberculosis, and

malaria combined. However, the number of RTI injuries

is more jarring, with 20-50 injured due to RTIs for every

RTI related death[1]. These 60 million injured patients expected

in the next ten years place a significant burden on

local, unequipped medical facilities. Regional economies

are also significantly impacted, with much of the local productivity

reliant on human labor.

1.1.2 – Rural-Medical Practitioners and Under-Qualification

This problem of exorbitant amounts of injuries is compounded

by the lack of qualified specialists and an excessive

number of under-qualified physicians referred to as

Rural Medical Practitioners (RMPs) who have no formal

education. These RMPs tend to over prescribe painkillers

in place of legitimate procedures, taking advantage of their

uneducated clientele. Due to the inaccessibility of expensive

equipment such as X-ray imaging devices, RMPs use

subjective and often inaccurate methods to track the rehabilitation

of patients with lower-extremity fractures. This

research aims to connect underserved communities with

more educated specialists, who are often located in larger

metropolitan areas.

Figure 1. Graph depicting concentrations of Road

Traffic Deaths across the globe [15]

1.1.3 – Delayed-Union and Non-Union

Approximately 5% to 10% of fractures worldwide proceed

to nonunion [4], which is the improper bonding of a

ENGINEERING

Broad Street Scientific | 2019-2020 | 51


fracture (Fig. 2). This leads to permanent disabilities and a

significantly higher need for healthcare resources, which

often can not be provided. Classically, the reasons for delayed-union

and non-union are complications including

inadequate reduction, loss of blood supply, and infection,

all of which are extremely prevalent in developing areas

[6]. A well designed post-operative rehabilitation protocol

can be implemented to identify any stagnations in the

healing process. However, physicians in rural areas often

cannot provide quality postoperative care, leaving their

patients without outpatient facilities after surgery.

1.2 – Background

1.2.1 – Telemetric Medicine

Telemetric medicine is a budding subfield of biomedical

engineering that allows for patient-doctor communication

from thousands of miles away. This has massive potential

in addressing the problems faced by underserved communities,

including a lack of qualified specialists, by giving

them access to these specialists. Currently telemetric medicine

has been limited to first world countries, where elderly

and disabled patients who are not capable of traveling to

a hospital can be provided with quality care from the comfort

of their home. However, current trends have shown

rapid increases in the number of smartphones available to

rural populations in developing countries due to a decline

in phone price. This has allowed for fast data transmission

to these areas. For example, vaccine reminders and natural

disaster alerts have become commonplace. This research

takes advantage of this framework by allowing for direct

patient to doctor communication over long distances.

1.2.2 – Weight-Bearing Ability

Weight-bearing ability of the afflicted limb has been

shown to greatly correlate with radiological evidence of

fracture union and overall healing (Fig. 3) [4]. According

to study by S. Morshed, 84% of patients indicated that

weight bearing ability was the most important clinical criteria

for diagnosis of delayed union and non-union. Despite

weight-bearing ability being the most critical factor

in the diagnosis of non-union, it is not often used due to

the subjectivity of clinical observation and patient feedback

being very unreliable and incomparable over the long

term. This research presents an objective method to track

gait parameters, such as weight bearing ability, and allows

for repeatability- crucial in the tracking of a patient’s rehabilitation

protocol.

Figure 2. X-Ray of a Humeral Nonunion [16]

1.1.4 – Benefits of Early Partial Weight Bearing

Early weight bearing is often absent from many rural

rehabilitation protocols, with “just rest” being prescribed,

as that is mistakenly seen as the best way to heal a fracture.

However, early partial weight-bearing has been found to

improve fracture healing, maintain bone stock and density,

and keep the fracture and implants aligned early in the

recovery process [5].

Figure 3. Graph depicting relationship between

weight bearing and fracture healing [17]

1.2.3 – Review of Insole-Based Gait Analytics Systems

Currently a majority of gait analysis is performed in

two specific methods: in a laboratory with force plates and

3-D motion tracking or in a doctor’s office with a clinician

making visual observations. The first method is extremely

expensive and provides highly specialized and unnecessary

data points while the second method is very subjective and

is often not repeatable over several trials and patient visits.

The benefits of insole-based gait analytics have been

identified as an accurate and cheap alternative to laborato-

52 | 2019-2020 | Broad Street Scientific ENGINEERING


ry and clinician-based gait analysis. Gait analysis systems

measure a range of data types including foot angle, stride

distance, step distance, step count, cadence, and speed . To

measure the extremely different statistics that fall under

the gait analysis the insoles use different sensor types for

each statistic. Accelerometers are used for stride length and

velocity, gyroscopes are used for orientation, flex-sensors

are used for plantar and dorsi-flexion, electric field sensors

for height above the floor, and finally force-sensitive resistors

are used for force parameters which is the focus of

this work [2].

1.2.4 – Ground Reaction Force and Biomechanics

The method utilized to measure weight bearing ability

is the temporospatial gait parameter known as vertical

ground reaction force (VGRF). In biomechanics, it is defined

as the force exerted by the ground on the body, when

in contact with the ground. Research has also suggested

that ground reaction forces correlate well with callus mineralization

and weight-bearing ability [3]. An increase in

ground reaction force indicates increased healing for the

patient. While weight-bearing is subjective, ground reaction

force measurements give healthcare providers a

singular statistic that can be stored and automatically processed

by computers in a setting where human involvement

is scarce and valuable. The VGRF graph (Fig. 4) displays

the active loading peak, the point of greatest force in

the gait cycle, which is the target statistic our insole aims

to measure.

plates. Force plates are traditionally used in gait analysis

but are not viable as even a slight increase in cost can exponentially

decrease the number of viable users. The insole

was also designed to be less than 300 grams, as any heavier

weight has been found to greatly alter gait parameters,

resulting in inaccurate data [2]. Small, individual sensors

were used in place of larger sensors, minimizing hysteresis

and inaccuracy in measurements across multiple trials.

2.2 – Force-Sensitive Resistors and Placement

The insole uses 4 force-sensitive resistors (FSRs) that

utilize polymer-thick film technology and provide a resistance

differential based on the force exerted. Polymer-thick

film sensors are constructed by the deposition of several

dielectric layers via a screen-printing process, making

them extremely cheap and lightweight compared to capacitive

sensors. To maximize cost efficiency, 4 locations on

the sole of the foot, which were identified via a pedobarograph

(Fig. 5) to have the greatest concentration of force,

are where the FSRs would be placed. The locations on the

sole of the foot determined to have the greatest concentrations

of force were under the big toe, metatarsal head I,

metatarsal head V, and the heel [11].

Figure 4. Vertical Ground Reaction Force graph over

a single stride [18]

2. – Materials and Insole Design

2.1 – Insole Design Criteria

The insole was chosen as the primary data acquisition

device due to the low relative cost in comparison to force

Figure 5. Pedobarograph representing areas of highest

force concentrations [19]

ENGINEERING

Broad Street Scientific | 2019-2020 | 53


2.3 – Hardware Structure and Data Pathway

The data from the FSRs are then sent to an ankle-mounted

Arduino Nano microcontroller attached via

a Velcro strap. The Arduino microcontroller compiles the

data received and uses a Bluetooth module to send the data

via Bluetooth low energy signals (Fig. 6). In gait analysis,

wireless technology is heavily utilized and beneficial as

wired technology greatly affects the target gait parameters,

resulting in non-representative data. The data are then received

by the data analytics software, which is written in

the Python programming language.

Table 1. Consecutive VGRF Values (Newtons) Acquired

from Insole (Left→Right and Top→Bottom)

0 0 0 0

203.2 364.8 389.6 406.4

429.2 403.2 324.8 194.4

0 0 0 0

0 0 0 0

183.2 348.8 367.2 373.6

388.4 364.8 0 0

0 0 0 0

208.8 336.0 365.6 372.8

393.6 404 212.8 0

0 0 0 0

Figure 6. Insole hardware wiring including an Arduino

Nano and HC-O6 Bluetooth Module

3. Data Analysis and Software

3.1 – Python Code and Peak Analysis

The Python program performs time-series peak analysis

to acquire the active loading peak of the ground reaction

force. Table 1 is an example of a raw dataset (in

Newtons) that requires peak analysis. Each data point in

Table 1 is taken every 30 milliseconds from the insole and

is stored in the table. Figure 7 is a plot of the data in Table

1 and depicts how peak analysis functions. The red point

in each step represents the active loading peak or the max

VGRF for that step. An algorithm included within the

code performs this process by defining a step as an array

of non-zero digits separated by zeroes, where the largest

value in that array is extracted and tagged as the VGRF of

that individual step. The non-zero digits are the instances

when the foot is in contact with the ground, while the zero

digits are when the foot is in the air, providing no data to

the force sensors. To provide robust and accurate data, the

program is designed to average the VGRF value over multiple

strides (N>6) during the trial. Six strides were chosen

as the minimum number of steps needed to provide representative

gait analysis data including ground reaction force

[7].

3.2 – Google Sheets API

The average VGRF value is then printed to a Google

Sheets API, allowing it to be accessed from any device. The

Google Sheets API provides the healthcare provider with

the week number relative to the beginning of the rehabil-

Figure 7. Peak Analysis Visualization

itation protocol, VGRF values, a visual representation via

a graph that is updated after every entry from the patient,

and an automatic risk assessment. As this work aims to

address the lack of qualified specialists, the API automates

the tedious aspect of screening at-risk patients during

rehabilitation. To achieve this, the API tags any patients

with a stagnation or reversal of VGRF as high risk, and

tags patients with increases in VGRF as low risk. Using

this novel approach, the healthcare provider can address

patients with a high risk of delayed or nonunion quickly

and accurately.

3.3 – Smartphone Application

To provide patient-centered care, a mobile app that

provides patients with data and communication tools for

proper management of their injury was developed (Fig. 8)

through the MIT App Inventor program. The application

includes access to the cloud database, a graph of weekly

values, and emailing functionality. The app also includes

Bluetooth pairing capability which allows the patient to

begin a trial through their smartphone. The design is focused

on simplicity of use. Upon initialization of the app,

the user is shown a welcome screen in which they can visit

their history or begin a test. Patient history includes a table

54 | 2019-2020 | Broad Street Scientific ENGINEERING


and graph of weekly data to aid patients in visualization

of the healing progress of their fracture. Testing is done

over individual IP servers in order to protect privacy and

connect with the API database.

hysteresis for the optimized and designed insole. A 2-Sample

T-Test for the final hysteresis values was performed

with 30 trials, resulting in a significance value of a=.05.

The test results in a p-value of 0.021 indicating statistical

significance.

5. Conclusion and Future Work

Figure 8. Mobile Application Initialization Pathway

4. Results and Discussion

4.1 – Hysteresis and Quality Analysis

FSRs are rarely used in biomedical devices, as consistency

and accuracy are one of the most important criteria

for success, but FSRs have classically been more suscep

tible to inaccuracies. An experiment was designed to test

the quality of the insole, the extent of hysteresis and the

inaccuracy of the sensor over time.

5.1 – Conclusion and Real-World Application

This work addresses a commonly overlooked lapse in

the healthcare system of a significant portion of the world’s

population. The addressed issue of increased road traffic

injuries will continue to exponentially grow in coming

years. Novel solutions to the proper healthcare and rehabilitation

of these injuries will be required to maintain a

high quality of life for those afflicted. The hardware utilized

is optimized for the specific use-case of a rural and

low-income setting in conjunction with a software application

designed to utilize the overdeveloped sector of

communications technology in developing countries with

many rural villagers owning smartphones.

5.2 – Future Work

5.2.1 – Machine Learning

Machine learning is another budding field of technology

that is uprooting the medical sector by automating

tasks that could previously only be performed by a trained

physician. However, a drawback of machine learning and

artificial neural networks is the vast amount of data needed

to train and create an accurate model [14]. The data that

are collected in the field from this work could contribute

to gait analysis databases and allow for even more automation

in this field, thereby improving the quality of life of

underserved communities in developing countries.

5.2.2 – Other gait disorders

Research has shown that other common disorders such

as Parkinson’s Disease and Cerebral Palsy have clear lower-extremity

gait related symptoms [12][13]. As the modern

hospital moves more towards the patient’s home, telemetric

medicine will become heavily utilized with medical

devices such as the one presented in this work.

Figure 9. Hysteresis or inaccuracy overtime: optimized

vs unoptimized sensor

Hysteresis or deviation is the inaccuracy of a sensor’s

indicated values from actual values which occurs in all sensors

and is accounted for. Figure 9 plots sensor hysteresis

against hours of use under the stress from weight that simulates

human locomotion. The control was a single unoptimized

FSR, and it was compared to the insole designed

with FSRs optimized by placement, size, and resistance.

The graph below shows clear reductions in the amount of

ENGINEERING

6. Acknowledgments

I would like to express my gratitude towards Dr. Sambit

Bhattacharya of Fayetteville State University for guiding

me in this project. His input allowed me to find solutions

in ways I never would have thought to.

7. References

[1] Mathew, G., & Hanson, B. P. (2009). Global burden of

trauma: Need for effective fracture therapies. Indian journal

of orthopaedics, 43(2), 111.

Broad Street Scientific | 2019-2020 | 55


[2] Bamberg, S. J. M., Benbasat, A. Y., Scarborough, D. M.,

Krebs, D. E., & Paradiso, J. A. (2008). Gait analysis using a

shoe-integrated wireless sensor system. IEEE transactions

on information technology in biomedicine, 12(4), 413-

423.

[3] Seebeck, P., Thompson, M. S., Parwani, A., Taylor, W.

R., Schell, H., & Duda, G. N. (2005). Gait evaluation: a tool

to monitor bone healing?. Clinical Biomechanics, 20(9),

883-891.

[4] Morshed, S. (2014). Current options for determining

fracture union. Advances in medicine, 2014

[5] Dorson, Jill R. "Biofeedback aids in resolving the paradox

of weight-bearing." (2018).

[6] Nunamaker, D. M., Rhinelander, F. W., & Heppenstall,

R. B. (1985). Delayed union, nonunion, and malunion.

Textbook of Small Animal Orthopaedics, 38.

[15] Naboureh, A., Feizizadeh, B., Naboureh, A., Bian, J.,

Blaschke, T., Ghorbanzadeh, O., & Moharrami, M. (2019).

Traffic accident spatial simulation modeling for planning

of road emergency services. ISPRS International Journal

of Geo-Information, 8(9), 371.

[16] Edginton, J., & Taylor, B. (2019). Humeral Shaft Nonunion.

OrthoBullets

[17] Joslin, C. C., Eastaugh-Waring, S. J., Hardy, J. R. W.,

& Cunningham, J. L. (2008). Weight bearing after tibial

fracture as a guide to healing. Clinical biomechanics, 23(3),

329-333.

[18] Larson, P.(2011). Vertical Impact Loading Rate in

Running: Linkages to Running Injury Risk.

[19] Tekscan, F-Scan System. https://www.tekscan.com/

products-solutions/systems/f-scan-system.

[7] Falconer, J., & Hayes, P. W. (1991). A simple method

to measure gait for use in arthritis clinical research. Arthritis

& Rheumatism: Official Journal of the American

College of Rheumatology, 4(1), 52-57.

[8] Rao, U. P., & Rao, N. S. S. (2017). The rural medical

practitioner of India. Evol Med Dent Sci, 6.

[9] Naeem, Z. (2010). Road traffic injuries–changing

trend?. International journal of health sciences, 4(2)

[10] Bachani, A., Peden, M. M., Gururaj, G., Norton, R., &

Hyder, A. A. (2017). Road traffic injuries.

[11] Razak, A., Hadi, A., Zayegh, A., Begg, R. K., & Wahab,

Y. (2012). Foot plantar pressure measurement system:

A review. Sensors, 12(7), 9884-9912

[12] Zhou, J., Butler, E. E., & Rose, J. (2017). Neurologic

correlates of gait abnormalities in cerebral palsy: implications

for treatment. Frontiers in human neuroscience, 11,

103

[13] Hausdorff, J. M. (2009). Gait dynamics in Parkinson’s

disease: common and distinct behavior among stride

length, gait variability, and fractal-like scaling. Chaos:

An Interdisciplinary Journal of Nonlinear Science, 19(2),

026113

[14] Gal, Y., Islam, R., & Ghahramani, Z. (2017). Deep

bayesian active learning with image data. In Proceedings

of the 34th International Conference on Machine Learning-Volume

70 (pp. 1183-1192). JMLR. org.

56 | 2019-2020 | Broad Street Scientific ENGINEERING


DIFFERENCES IN RELIABILITY AND

PREDICTABILITY OF HARVESTED ENERGY FROM

BATTERY-LESS INTERMITTENTLY POWERED

SYSTEMS

Nithya Sampath

Abstract

Solar and radio frequency (RF) harvesters serve as viable alternative energy sources for battery-powered devices in which

the battery is not easily accessible. However, energy harvesters do not consistently produce enough energy to sustain an

energy consumer; thus, both the energy availability and execution of the consumer process are intermittent. By simulating

intermittent systems with large-scale energy demands using specifically designed circuit models, parameters including

harvested voltage, voltage across the capacitor, and voltage across the consumer were determined. The probability

of energy availability was then computed based on the number of energy events that had previously occurred for both

harvested solar and radio frequency. A metric designated as the η-factor was computed from the probability plots for the

solar and radio frequency data to quantify the reliability of each power source. The η-factor for harvested solar energy

was significantly higher than that of harvested radio frequency energy, indicating that harvested solar energy is more

consistently available than harvested radio frequency energy. Finally, the effects of various obstacles between the radio

frequency transmitter and receiver on the output voltage were determined. Increasing the distance between the transmitter

and receiver, as well as placing people and metal between the two, resulted in a significant drop in energy availability as

compared to foam and wood obstacles. Quantifying the reliability of different harvested sources would help in identifying

the most practical and efficient forms of renewable energy. Determining which obstacles cause the most obstruction to a

signal can aid in optimizing the strategic placement of harvesters for maximum energy efficiency.

1. Introduction

Battery-powered devices such as pacemakers and neurostimulators

are not suitable in many systems due to their

inaccessible placement and frequent need for battery replacement.

Harvested energy from solar, radio frequency,

and heat sources is an attractive alternative to batteries

in such systems; however, energy availability from these

sources is not consistent and is therefore characterized as

intermittent. There are two components to these intermittently

powered systems: the energy harvester and the energy

consumer. The energy consumer utilizes the energy

captured by the energy harvester. For the energy consumer

to remain functional, a certain power threshold must be

met. There is often a discrepancy between the amount of

energy required to power the consumer and the amount

of energy supplied to the system by the energy harvester,

causing the device to cyclically turn on and off as shown

in Figure 1. This sporadic energy harvesting pattern leads

to an interrupted, or intermittent, execution of the energy

consumer [1]. Most renewable energy sources, including

solar and wind power plants, have an intermittent power

output [2].

An energy event is defined as the generation of a significant

voltage in a given time interval. In this study, a

minimum voltage of 2.8V during a period of 5 minutes was

considered significant. Indicating that N energy events

have occurred signifies that 5N minutes have passed since

the energy consumer last shut off. The time designated for

an energy event in this investigation was selected to test

the burstiness of energy and accurately determine when

the probability of a future energy event reaches 0. A capacitor

was included in the circuit to store and release energy

slowly into the consumer. Without the capacitor, energy

flowing directly from the harvester to the consumer would

instantly terminate the execution of the consumer software

in the absence of an energy event [3].

Figure 1. Graph of voltage vs. time for an energy consuming

device. The device turns on when it reaches

a threshold voltage, approximately 2 Volts, and begins

to consume energy at a rate greater than the rate

at which harvested energy is supplied to the circuit.

When the voltage supplied to the energy consuming

device drops, the device shuts off and the voltage is

allowed to increase again from the harvested power

supply. This cycle repeats.

ENGINEERING

Broad Street Scientific | 2019-2020 | 57


The objective of this study is to model an energy harvesting

system on a small scale using solar harvester and

radio frequency harvester units. This model will be used

to analyze the burstiness of energy by computing the likelihood

of power availability, given that N consecutive energy

events have occurred. It is hypothesized that all harvested

energy will be available in short bursts, consistent

over small periods of time. The reliability of various harvested

sources in terms of their power availability will also

be determined by computing the η-factor. The η-factor is

a metric that compares the experimental energy harvesters

to a random energy harvester, in which energy events are

independent. This is not the case for real energy harvesters,

in which energy events are dependent upon each other.

It is predicted that the η-factor of solar energy will be

higher than that of radio frequency energy, both of which

will be less reliable than wall power, which has a η-factor

of 1.0. Lastly, the effect of various obstacles and distances

between an RF transmitter and receiver will be determined.

It is expected that obstacles with a higher density,

such as people, will allow for less energy to be harvested

than obstacles such as foam or wood. Obstacles were considered

only for harvested RF energy, as solar energy is

traditionally harvested without obstruction. This research

currently has great relevance as the world begins to look

towards renewable energy sources to replace fossil fuels.

Investigating the patterns of energy availability and consumption

allows for precise prediction of energy patterns

and optimization of execution process scheduling.

2. Materials and Methods

The setup for both the solar and the radio frequency experiments

involved two main circuits as shown in Figure

2. The first, designated as the harvester circuit, captured

energy either from the sun or from the transmitted RF

signal and converted it into electrical energy. There were

4 main components to this circuit: the solar or radio frequency

harvester unit, the capacitor, the LTC (Load Tap

Changer), and the consumer. In this experiment, the load

was an MSP-430 device that was constantly running an

energy-consuming software and required 2.8 Volts to turn

on. The purpose of the LTC in the circuit was to ensure

that given a certain amount of energy input, the amount

of voltage output would be 3.3 Volts. This amplified input

voltage pronounces the absence of sufficient energy.

The voltage was recorded across the energy consumer and

energy harvester, which captured energy from the sun

or radio frequency transmitter. The weather conditions

during the experiment are important to consider when

dealing with solar data. A light sensor was used to record

the full-spectrum, infrared, and visible light levels. The device

also retrieved real-time weather data from Weather.

com, including the temperature and UV index. The experiment

was conducted under consistent weather conditions,

sunny and minimal cloud coverage, and the harvester was

placed directly across from a window facing west. The system

tended to reach peak energy in the afternoon, when

the most sunlight was available.

Figure 2. Diagram of circuit setups for solar and radio

frequency harvesting experiments. Output voltages

from each component in the harvester circuit — the

harvester unit, the capacitor, the LTC, and the load

— were connected to the Arduino and recorded.

For the radio frequency experiment, an infrared sensor

was also used to record the presence and absence of people;

a person walking in front of the sensor and blocking the

signal was considered as an absence of an energy event. A

secondary logger circuit was constructed to record voltages

at each point in the harvester circuit. It consisted of a

Raspberry Pi 3 and an Arduino Uno, powered from the

wall. Arduino and Python programs were used to record

and analyze the voltage data, calculate η-factors, and construct

plots of energy availability.

The obstacle experiment with a radio frequency harvester

involved a similar setup, except for an infrared sensor.

Python and Arduino programs were used to record the

voltages at various points in the circuit. Data were collected

over a span of 2 minutes each for four different obstacles–metal,

wood, person, and foam–over three different

distances – 1 meter, 2 meters, and 3 meters. The voltages

were recorded from the D out

pin on the MSP-430, which

is directly related to the amount of radio frequency input

received. The absence of an obstacle between the transmitter

and the receiver served as a control to which the other

voltage values were compared. The approximate densities

and thicknesses of the obstacles used are listed in Table

1. The density of a human was approximated to be 1.01

grams per cubic centimeter [4].

3. Results

This study aimed to test the burstiness properties of

harvested energy by analyzing the probability of energy

availability. The idea that energy was available in short

58 | 2019-2020 | Broad Street Scientific ENGINEERING


Table 1. Approximate thicknesses and densities of

obstacles used in experiment. The thicknesses were

measured and densities were calculated using measurements

of mass and volume. The thickness measurement

for a human is very approximate as it is

harder to measure with precision.

Approximate Thickness and Density of Obstacles

ENGINEERING

Thickness (cm) Density (g/cm 3 )

Human 25 1.01

Metal Whiteboard 2.0 0.85

Wood 2.0 0.70

Foam 0.3 0.13

bursts was suggested by earlier data collected in the lab;

however, the data collection did not occur for long enough

to prove conclusions supporting that idea. To determine

whether the energy-consuming device was turned on and

how much energy was being supplied to the circuit, voltage

data was recorded at various locations in the circuit. In

Figure 3 (solar) and Figure 4 (radio frequency), the horizontal

axes contain both positive and negative values for

the number of previous energy events. Negative numbers

of energy events correspond to a continuous absence of

energy events. These graphs support the hypothesis of the

burstiness of harvested energy since the probability does

not oscillate as the number of energy events increase.

Figure 3. Probability plot for harvested solar energy.

This plot displays energy occurring in short bursts.

The mean η-factor is 0.8595 (n = 6). The data collection

for this experiment spanned three days.

Figure 3 also depicts that the correlation between the

probability and the number of energy events decreases

near N = 70. This is consistent with the designated definition

of an energy event, as 70 energy events lasting 5

minutes each would cumulate to 5.83 hours. This is the

approximate time for which sufficient sunlight was facing

the energy consumer in the experiment location. This feature

is not represented in Figure 4 as the amount of harvested

radio frequency energy was not dependent upon the

time of day.

Figure 4. Probability plot for harvested radio frequency

energy. This plot displays energy occurring

in short bursts. The mean η-factor is 0.3657 (n = 6).

The data collection for this experiment spanned two

weeks.

The reliability of various harvested sources based on

the η-factor was compared. The findings reinforce the

hypothesis of a given harvested energy source being more

reliable than another due to the variance in reliability patterns.

The mean η-factor for harvested solar power was

0.8595 and the mean -factor for harvested radio frequency

power was 0.3657. The standard deviation of the solar data

was 0.0018 and the standard deviation of the radio frequency

data was 0.0762. A student’s t-test was performed

over the 6 trials for each of the solar and radio frequency

experiments. The t-test yielded a two-tailed p-value of less

than 0.0001, indicating a statistically significant difference

between the η-factors for harvested solar and harvested

radio frequency energy. The η-factors for both harvested

solar and harvested radio frequency energy fall below the

ideal standard of 1.0, the η-factor for wall and battery power.

This can be observed in Figure 5, where the probability

of energy availability, given that any number of energy

events have occurred, is 1.0 [5]. The higher η-factor for

solar power suggests that harvested solar energy is more

reliable than harvested radio frequency.

Finally, various obstacles and distances separated a radio

frequency transmitter and receiver to determine how

they would affect the amount of harvested energy. Varying

obstacles are likely to have varied effects on the amount of

energy able to be harvested, due to the density and thickness

of the object. The results shown in Figure 6 indicate

that metal and people had a more pronounced effect on the

ability of the receiver to harvest energy from the transmitter

as compared to wood and foam. Foam had a slightly

higher voltage input value than the absence of an obstacle

at 1 meter, but this difference is too small to be significant

and was likely caused by random variation. Additionally,

Broad Street Scientific | 2019-2020 | 59


as distance increases, the received signal input decreases

for all obstacles. Over larger distances, fewer signals can be

received and converted into electrical energy.

Figure 5. Theoretical probability plot for wall power.

Wall power has a η-factor of 1.0 since it is not intermittent.

Figure 6. Graph depicting D out

for various obstacles

and distances in the RF obstacle experiment. The D out

value is correlated with the radio frequency input,

so higher D out

values correspond to higher radio frequency

input.

4. Discussion

The major objectives for this investigation included

investigating the burstiness of energy, comparing the reliability

of harvested solar and harvested radio frequency

energy, and exploring the effects of various obstacles and

distances on the amount of harvested radio frequency energy.

Graphs of the probability of energy availability given

that a certain number of energy events had occurred were

constructed and analyzed from which η-factors were calculated

(Fig. 3 and 4). The average D out

values for different

obstacles and distances were also plotted against a control

group, comparing which obstacles had the greatest effect

on D out

values (Fig. 6).

Figures 3 and 4 indicate the burstiness of both harvested

solar and radio frequency energy. The probability of an energy

event occurring becomes relatively high, after many

have already occurred. This correlation does not continue

in the case of harvested solar energy. Increasing the time

interval t, which defines an energy event, could have had

an influence on this data. A larger value of t may represent

sections of the graph after the correlation ends for solar

energy, causing probabilities in the middle to appear closer

to 1 than they truly are. Selecting a smaller value of t

may not show where the correlation ends, suggesting that

the correlation does not in fact end. In Figure 4, there is

a spike in probability at approximately N=-1. This indicates

that if there was no energy event in the previous time

interval, then the probability of a future energy event is

extremely high. This suggests that when a person walks in

front of the sensor, they cause the absence of a single energy

event, but they are not likely to cause the absence of a

second energy event. In other words, most people walk by

the sensor rather than standing in front of it. Inaccuracies

in the infrared sensor, indicating if a person was blocking

the signal, may have also influenced the results. Finally, the

Arduino could only record voltages to 2 decimal places,

limiting the precision of the data analysis. This analysis of

the burstiness of energy will allow for the optimization of

scheduling execution processes.

The calculations of the η-factor and the results of the

student’s t-test suggest that harvested solar power is more

reliable than harvested radio frequency power, supporting

the original hypothesis. The increased standard deviation

for the radio frequency-factor could be attributed to the

increased variance among trials in the patterns of people

passing in front of a sensor over a given time, compared

to the more stable pattern of light reaching a solar panel.

Data collection for the harvested radio frequency energy,

which lasted 2 weeks, spanned a longer period than that of

solar data, which lasted 3 days. The rationale behind this

was that the absence of radio frequency energy events is

less frequent than the absence of solar energy events, so

more data are needed to gain a holistic understanding of

the radio frequency energy patterns. As in the first experiment,

the reliability of the sensor and Arduino device

could have also influenced the data collected. These results

indicate that harvested solar energy may be a more suitable

alternative to harvested radio frequency energy in terms of

reliability and predictability.

The results presented in Figure 6 suggest that people

and metal, more than other obstacles, obstruct the radio

frequency signal from being received and converted into

electrical energy. Thus, people were used to obstruct the

signal in the first radio frequency experiment in order to

induce the absence of an energy event. Factors that may

have influenced the results include the presence of multiple

objects between the transmitter and receiver, such

as a foam board being held by a hand. These results are

relevant to the real-world application of harvested radio

frequency energy, including Wi-Fi routers and cell towers.

In addition, the results indicate which obstacles are more

likely to cause the absence of an energy event and can in-

60 | 2019-2020 | Broad Street Scientific ENGINEERING


form the placement of radio frequency energy harvesters.

There were limitations in some aspects of the model

regarding accurately modeling large-scale solar and radio

frequency energy harvesting systems. For the radio frequency

experiment, the model fails to account for the real-life

conditions of multiple obstacles or static obstacles,

which could decrease the likelihood of energy events by

allowing for less energy to be successfully harvested.

Future research may involve testing different forms of

harvested energy under various weather conditions. The

differences in amounts and reliability of harvested energy

between different types of piezoelectric energy could be

investigated, including energy harvested from mechanical

stress inside a person’s shoe or energy harvested from mechanical

stress on a tile. As solar cell technology advances,

the reliability of different types of solar panels could also

be tested.

In the search for fossil fuel replacements in this era of

climate crisis, the reliability of different harvested energy

sources should be considered. Harvested solar energy was

found to be more reliable than harvested radio frequency

energy, and this should be a factor in deciding which

types of renewable energy to invest in and implement on a

large scale. These harvesters, particularly radio frequency

harvesters, should be placed in a way that minimizes obstruction,

particularly from people and metal. In this way,

energy efficiency can be maximized. Through harvesting

energy from common and everyday sources like the sun

and radio frequency signals, the first steps are taken towards

ensuring a more sustainable and energy-efficient

future.

[3] Islam, B., Luo, Y., Lee, S., & Nirjon, S. (2019, April).

On-device training from sensor data on batteryless platforms.

In Proceedings of the 18th International Conference

on Information Processing in Sensor Networks.

[4] Density of Human body in 285 units and reference

information (2019). Retrieved November 9, 2019.

[5] Islam, B., Luo, Y., & Nirjon, S. (2019). Zygarde:

Time-Sensitive On-Device Deep Intelligence on Intermittently-Powered

Systems. arXiv preprint arXiv:1905.03854.

5. Acknowledgements

Thanks to Dr. Shahriar Nirjon and Bashima Islam of the

Department of Computer Science at UNC-Chapel Hill

and to Dr. Sarah Shoemaker, Ms. Shoshana Segal, and

Mr. Chris Thomas of the NCSSM Mentorship and Research

Program for advice and assistance throughout the

research process.

6. References

[1] Lucia, B., Balaji, V., Colin, A., Maeng, K., & Ruppel,

E. (2017). Intermittent computing: Challenges and opportunities.

In 2nd Summit on Advances in Programming

Languages (SNAPL 2017). Schloss Dagstuhl-Leibniz-Zentrum

fuer Informatik.

[2] Fuchs, E., & Masoum, M.A. (2015). Power Quality in

Power Systems and Electrical Machines (2nd ed.). Academic

Press Inc.

ENGINEERING

Broad Street Scientific | 2019-2020 | 61


APPLYING MACHINE LEARNING TO HEART DISEASE

DIAGNOSIS: CLASSIFICATION AND CORRELATION

ANALYSES

Sahil Pontula

Abstract

Classification machine learning has emerged to quickly and efficiently analyze large sets of data and make predictions

about potentially unknown variables. It has seen application in numerous fields, from sorting through ionosphere signals

to predicting weather. Here, we report on the application of classification algorithms to heart disease diagnosis, run on

a sample dataset of 303 patients. We test several models and compare their results to determine accuracy in predicting

presence or absence of heart disease. Furthermore, we conduct statistical and graphical analyses to determine correlations

and causations between different attributes linked to cardiovascular disease. We believe that the methods demonstrated

here show promise for large-scale applications, for both more complex and comprehensive datasets and real-time data

collected in a clinical setting.

1. Introduction

1.1 Heart Disease

1.1.1 Angina

Approximately 9.8 million Americans are thought to

suffer from angina pectoris, or simply angina, annually,

and it is suspected that 500,000 new cases emerge each year

[1]. Angina is a common symptom of heart attacks (myocardial

infarctions) and is the result of insufficient blood

supply to the heart (cardiac) muscle, known as ischemia.

In ischemia, the heart does not receive enough oxygen to

pump adequately. Despite angina’s relative commonness,

it can be difficult to distinguish between other causes of

chest pain. The chest pain of the patients whose data we

investigate here is classified as anginal or non-anginal (and

the cause may be known or unknown).

Symptoms of angina include chest pain, pain in the

limbs or jaw that accompanies chest pain, nausea, sweating,

fatigue, and shortness of breath, though symptoms for

females can be different from the conventional symptoms,

and as such treatments will vary. Angina can be classified

as stable or unstable. Stable angina is more common and

can occur during exercise (and disappear upon resting). It

occurs when the heart is required to pump much harder, is

short-lasting, and is often predictable. Unstable angina is

more dangerous and may indicate risk of a heart attack (it

is unpredictable and longer-lasting). A third type of angina,

variant or Prinzmetal’s angina, is rarer and is caused by

a spasm in the coronary arteries [2].

Most often, the ischemia that causes angina is due to

coronary artery disease. When coronary arteries are

blocked by fatty deposits, or plaques, the condition is

known as atherosclerosis. Exercise-induced angina occurs

when the effects of ischemia are felt during exercise (reduced

blood flow through coronary arteries cannot match

the oxygen demand).

Risk factors of angina include tobacco use, diabetes,

high blood pressure, high blood cholesterol levels, elderly

age, sedentary lifestyle, obesity, and stress. We investigate

just some of these factors in analyzing the given dataset.

1.1.2 Age

It is known that individuals of age 65 or older are more

susceptible to developing heart disease, including myocardial

infarctions (heart attacks) and heart failure. With older

age, the heart cannot beat as quickly during exercise or

stress, even though the heart rate is fairly constant. Aging

commonly causes the arteries to grow stiffer, resulting in

a condition known as arteriosclerosis. This, in turn, results

in complications such as hypertension. Other changes

in the cardiovascular system that result from aging include

arrhythmias (irregular heartbeats or abnormal heart

rhythms) and edema.

1.1.3 Gender

At younger ages, males are more vulnerable to heart disease

than women; on average, men suffer their first attack

at age 65, while women suffer their first attack at age 72

[3]. Nevertheless, heart disease is the most prevalent cause

of death in both genders. Many factors may contribute to

the greater vulnerability of males, including higher rates of

risk factors such as smoking and the fact that hormones in

younger women protect against heart disease. Additionally,

the symptoms may differ between the genders; some

researchers have noted that during heart attacks, women

are more likely to suffer from abnormal symptoms.

Ironically, the survival rate after developing heart disease

shows a trend opposite to that of acquiring it. Survival

rates have been found to be lower in women, perhaps because

women are less likely to obtain advice or beneficial

62 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


medications than men or because women are generally

older and suffer from other health complications [3].

1.1.4 Cholesterol

Cholesterol, in normal levels, is essential for many body

functions, including building new cell walls and making

steroid hormones. In excess amounts, however, cholesterol

tends to accumulate in arterial walls, resulting in a

condition known as atherosclerosis. This disease results in

narrowed arteries and blocked or slowed blood flow, causing

angina or heart attacks. Note that blood-borne cholesterol

is generally characterized as low-density lipoproteins

(LDLs) or high-density lipoproteins (HDLs). LDLs are

responsible for the atherosclerosis-causing fatty plaques.

HDLs are thought to function in clearing the blood of excess

cholesterol. Additionally, high levels of triglycerides,

another type of fat, are suspected to correlate with occurrences

of heart disease [4].

1.1.5 Electrocardiograms (ECGs)

ECGs provide a convenient way to measure the electrical

activity of the heart, and hence find common application

in diagnosing cardiac disorders. A representative signal

of normal cardiac electrical activity is shown in Figure

1. The P wave is associated with atrial depolarization, the

QRS complex is associated with ventricular depolarization

and atrial repolarization, and the T wave is associated with

ventricular repolarization. The measurements are recorded

from a set of electrodes placed on the skin and the results

are typically presented as a “12-lead ECG,” which has

12 voltage-time graphs.

The ST segment connects the points marked S and T

in Figure 1. ST depression, wherein the ST segment is unusually

low relative to the baseline, is sometimes associated

with heart disease. It is generally indicative of myocardial

ischemia, which is primarily caused by coronary artery

disease (CAD), in which blood flow is greatly reduced for

much the same reason as in atherosclerosis. ST depression

is also seen in patients with unstable angina. The ST segment

may be depressed but also upsloping, in which case

ischemia is generally not the cause. Instead, it is likely a

variant of the normal sinus rhythm. Additionally, studies

have shown that the ST segment slope may provide an accurate

diagnosis of CAD [5].

1.2 Classification Algorithms

Here we describe the algorithms (and other related concepts)

used in the classification analysis of the heart disease

dataset.

1.2.1 KNN Algorithm

The KNN, or k-nearest neighbors, algorithm (Fig. 2)

is part of supervised machine learning techniques (“supervised”

indicates that labeled input data is being used for

train- ing/testing). The algorithm assumes that data similarly

classified are also close together geometrically (usually

measured by the Euclidean distance).

Figure 1. A sample ECG signal for a normal (sinus)

rhythm, showing the P wave, QRS complex, and T

wave [6].

MATHEMATICS AND COMPUTER SCIENCE

Figure 2. The steps of the KNN classification algorithm.

Note how distances are computed for the new

data point between it and members of the preexisting

classes before a decision is made for its classification

[7].

For classification, it works by computing the distance between

a data point and all other data points, selecting the

k nearest neighbors thus found (where k is defined by the

Broad Street Scientific | 2019-2020 | 63


user), and classifying according to the most frequently occurring

label. However, this algorithm can slow significantly

as the size of the dataset grows. In addition, the user

must decide on the value of k. If k is too small, the dataset

could be overfitted, leading to significant variance, while

if k is too large, the data may be oversampled, resulting in

bias.

1.2.2 Decision Tree Analysis

Decision trees function in classification and regression

in machine learning. They are conventionally drawn such

that their root is at the top. Conditions, or internal nodes,

are places where the tree splits into branches (edges).

When the branch does not split any more, we have found

the decision (leaf). Decision trees allow relations and important

features to be easily seen.

Decision trees are often based off of the technique of

recursive binary splitting, where the split points are selected

based on a cost function. This allows selection of

the splits that maximize the tree’s accuracy. Sub-splits are

also chosen recursively in this manner, so recursive binary

splitting is also known as the greedy algorithm, as the only

goal is to minimize cost. This means the root node, at the

top of the tree, is the best classifier. Gini scores, G = ∑p k

(1

−p k

), provide a good way to quantify how good a split is by

measuring how independent or mixed the two resulting

groups are. An ideal split would have G = 0.

Note, however, that trees that are too long may overfit

the data, and as such the number of training inputs used

for splitting is often limited.

1.2.3 Support Vector Machine Analysis

This supervised learning algorithm’s objective is to

find a hyperplane in n-dimensional space, where n is the

number of features in the data, that classifies the data by

separating the classes as much as possible. Many planes

are possible to separate two given classes, but the one that

maximizes the distance between data of different classes,

or margin distance, is chosen.

Support vectors are data points that lie closer to the hyperplane

and affect its position, directly affecting the margin

distance. Note that the margin distance is optimized by

a loss function known as the hinge loss function.

1.2.4 Neural Network Analysis

Typically, neural networks (NN) are the preferred

method of choice for classification tasks such as this one.

These structures consist of an input layer, an output layer,

and a series of hidden layers in between. All layers may

consist of multiple neurons (or nodes), which take in some

input value, perform a computation using some algorithm,

and propagate an output value to the next neuron. In the

end, the output from the output layer determines the result.

For classification tasks, the NN uses training data to

adjust the “weight” of the algorithm/computation that the

neurons use, where it can correct the values of the “weight”

to match the outputs of the training data. Once trained, the

NN can be used on a test set, predicting the appropriate

classifications. The results can be compared with the true

classifications to determine the accuracy of the model, and

eventually the model can be used with real-world data.

1.2.5 Random Forest Analysis

Random forests consist of many individual decision

trees working as an ensemble. Each tree outputs a class

prediction, and whichever class is predicted the most by

the trees is chosen as the model’s prediction. The random

forest algorithm is one of the best among supervised classification

techniques, because it works by integrating the

predictions of many independent, uncorrelated models

(trees), compensating for the individual errors of any one

tree [8].

1.2.6 Naive Bayesian Analysis

This classification technique is based on Bayes’s theorem

for computing conditional probabilities. It assumes

independence among predictors/features in the data, and

uses Bayes’s theorem to predict the probability of a given

data point falling into a certain class based on the predictors.

Naive Bayes is a quick, efficient way to classify data with

multiple classes. Unfortunately, it is accurate only if the

assumption of independence between predictors holds.

Additionally, Naive Bayes is often better with categorical

variables, as it assumes numerical variables are normally

distributed. Nevertheless, the fast nature of the algorithm

makes it a preferred method for real-time and multi-class

prediction [9].

1.2.7 Gini Index

This index measures the probability of a specific variable

being wrongly classified when randomly chosen. G =

0 or G = 1 indicate that all elements are in only one class

or the elements are randomly distributed across the classes,

respectively. G = 0.5 indicates equal distribution of the

elements in the classes.

The Gini index is defined as G = 1 − ∑(p i

) 2 , with p i

being

the probability an element is classified in a specific class.

In decision trees, the feature with the lowest Gini index is

usually chosen as the root node [10].

1.3 Objective

In this investigation, we have two primary purposes.

First, we attempt to use statistical analysis in R/RStudio,

including correlograms and the Bayesian information

criterion, to identify potential correlations and causations

among the variables in the heart disease dataset, described

in more detail below. Second, we use classification machine

learning algorithms, also in R/RStudio, to predict

the presence of heart disease in arbitrary patients.

64 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


2. Datasets

The data for this work were obtained from https://

archive.ics.uci.edu/ml/datasets/heart+Disease. Data were

collected from 303 patients. The original dataset consists

of 75 variables, but we only consider a subset of the 14

most important variables (which were used in published

experiments). The last variable is the class variable, referring

to the presence or absence of heart disease. Values of

1, 2, 3, and 4 denote presence (according to a scheme used

by the creators of the dataset), while a value of 0 denotes

absence. The 14 variables in the dataset are as follows:

1. Age, numerical

2. Gender, categorical

3. Chest pain type, categorical

4. Resting BP (mmHg), numerical

5. Serum cholesterol concentration (mg/dl), numerical

6. Fasting blood sugar concentration (mg/dl), categorical

7. Resting EKG results, categorical

8. Maximum heart rate (bpm), numerical

9. Presence of exercise-induced angina, categorical

10. ST segment depression due to exercise, numerical

11. Slope of peak exercise ST segment, categorical

12. Number of major vessels colored by fluoroscopy,

categorical

13. Status of defect, categorical

14. Diagnosis of heart disease, categorical

3. Data Preparation and Modeling

In order to clean the data, data munging was performed

in R/RStudio. The raw data were read as a CSV file directly

from the URL above. The columns were then renamed

to readable forms and missing data were found and

corrected for. Six rows had missing data, and we chose to

simply delete these from the dataset. We then reclassified

the categorical variables, changing the numbers to more

readable text.

4. Correlation Analysis

MATHEMATICS AND COMPUTER SCIENCE

For preliminary analysis, pairwise histograms and correlations

(scatter plots and correlation coefficients) were

produced for the five numerical variables. From the histograms,

it was noted that most numerical variables showed

only slight skew and the only variable with considerable

skew (ST segment depression) could not be easily corrected

with normalization transformations. As such, the data

were not transformed in any way. Furthermore, a correlogram

was generated, taking the numerical variables in

pairs, to visualize any correlations present.

Linear regressions were then created with ST segment

depression as the dependent variable. A multiple regression

model was also developed to predict ST segment depression.

To examine causality, Bayesian information criterion

(BIC) values were generated between ST segment depression

and each of the other variables. If the value for the

association between ST segment depression and a given

variable was more than 10 less than the value with no coupling,

a causal relationship was possible.

Lastly, two scatterplots were generated with ST segment

depression as the dependent variable and maximum

heart rate as the independent variable. Points were

grouped by two categorical variables, namely slope of peak

exercise ST segment and type of chest pain.

5. Classification Analysis

After designating the diagnosis variable as the class variable,

Gini indices were generated for all other attributes.

Lower values indicated more important variables, namely

those that would give greater equality in the distribution of

data into classes. These variables, in turn, could be used as

root nodes in decision trees. The attributes with the lowest

Gini indices were plotted using jitterplots, where the

dependent variable was heart disease diagnosis, to examine

data clustering.

With these important variables identified, various classification

algorithms were then tested. Among these were

the random forest, partition, neural network, support vector

machine (SVM), and Naive Bayes models. These algorithms

were run on two of the “best” categorical variables

(according to the Gini coefficients) and the best numerical

variable, ST segment depression. k = 3 clusters were used

for the neural network model, based on the jitterplots. Tables

were generated to document the accuracy of the algorithms

in predicting presence or absence of heart disease

in the 303 patients.

Additionally, a k-nearest neighbors (KNN) algorithm

was run. 70% of the dataset was designated for training,

and the remaining 30% was used to test the model. A cross

table of the model’s predictions and the actual diagnoses

was then produced and compared with the results of the

other classification algorithms.

6. Results

6.1 Correlation Analysis

The pairwise correlations and histograms for the numerical

variables are shown in Figure 3. The correlogram

for the numerical data are shown in Figure 4. Note from

both figures that most numerical data are distributed over

a range of values (i.e. more continuous than discrete). We

do not see particularly strong correlations, but this does

not rule out associations between the variables, as considerable

scatter may be present in the data.

Linear regressions between ST segment depression and

each of the other numerical variables are shown in Figure

Broad Street Scientific | 2019-2020 | 65


5. Again, we see no strong correlations. At discrete values

of age and resting blood pressure, we note that the data

points take on a range of ST segment depressions. From

the multiple regression plot shown in Figure 6, we again

see considerable scatter. Indeed, when the actual ST segment

depression is 0, the predicted depression adopts a

wide range of values.

Figure 5. Four linear regressions with ST segment

depression as the dependent variable. Age (years),

maximum heart rate (bpm), resting blood pressure

(mmHg), and blood cholesterol concentration (mg/

dl) are the independent variables.

Figure 3. Pairwise histograms and correlations for

the numerical variables. Note how the age histogram

is fairly normally distributed. The histograms

for resting BP and cholesterol level show right skew,

the histogram for maximum HR shows left skew,

and the histogram for ST depression shows significant

right skew. No strong correlations appear to be

present, the strongest being between age and maximum

HR.

Figure 6. A plot of the actual ST segment depression

and the predicted depression using a multiple regression

model with all four other numerical variables.

Figure 4. A correlogram for the numerical data in the

heart disease dataset. The results shown confirm the

findings from the pairwise plots. No strong correlation

are apparent. Note that the color indicates the

the direction of the correlation (positive, negative,

or no correlation) and the size of the circle indicates

the relative strength of the correlation.

Conducting BIC analyses with ST segment depression

as the dependent variable, it was found that half of the

variables had a possible causal relationship. Among these

were chest pain type, maximum heart rate, presence or absence

of exercise-induced angina, slope of the peak exercise

ST segment, and presence or absence of heart disease.

The latter two had the lowest BIC values.

Figure 7 shows plots with ST segment depression and

maximum heart rate, grouped by two categorical variables.

All variables showed evidence of correlation via the regressions

or BIC analyses. Grouping by slope, we see that flat

and downsloping ST segments show considerable scatter,

with the downsloping datapoints suggesting potential outliers

at high depressions. The upsloping ST segments seem

to be concentrated at low depressions. A similar trend is

seen in the second graph, where asymptomatic chest pains

suggest outliers. Nevertheless, none of the four categories

shows distinctive patterns in the plot.

66 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


algorithms. Similar accuracies among many of the algorithms

can be explained given that the dataset is not too

complex, so for most data points, the algorithms operate

similarly and give the same predictions. The accuracy of

the KNN algorithm, using all non-class variables as predictors,

was 64.9% for absence and 46.3% for presence of heart

disease (using the test dataset). We see that the partition

model using ST segment depression as a predictor has the

best accuracy for both absence and presence.

Figure 7. Two scatter plots with ST segment depression

as the dependent variable. The data points are

grouped by the categorical variables chest pain type

and slope of ST segment during peak exercise.

6.2 Classification Analysis

By finding Gini values for each of the non-class variables,

it was found that gender, fasting blood sugar concentration,

and presence/absence of exercise-induced angina

were most important. None of the numerical variables

were major predictors of diagnosis. Nevertheless, we show

jitterplots for the three aforementioned categorical variables

as well as the numerical variable with the lowest Gini

index in Figure 8. The three plots for categorical variables

show four distinct clusters, as expected. The plot with fasting

blood sugar (FBS) shows a concentration of data at low

FBS values, while the plot with exercise-induced angina

shows that most patients lacked both it and heart disease.

In the plot with ST segment depression (the numerical

variable), it is apparent that most patients presented with

low ST segment depression (especially those without heart

disease), with a few outliers at high depression values.

Figure 8. Four jitterplots for the variables with the

lowest Gini indices. Note the distinct clustering tendency

of the datapoints, indicating that most of these

variables are good clustering attributes and likely

important parts of any classification algorithm.

Table 1 shows the results of five of the classification

Table 1. The results of the classification algorithms,

excluding KNN. The first and second numbers in

each cell indicate the accuracy in predicting absence

or presence of heart disease, respectively. The partition

model appears to have the optimal accuracy in

predicting both absence and presence, using ST segment

depression as a predictor. Factors were chosen

based on Gini indices and the correlation analysis.

Naive

Bayes

ECG Result

(%)

Exercise-

Induced

Angina (%)

ST Segment

Depression

(%)

58.1 57.7 85.6 54.0 86.9 46.7

SVM 57.5 59.9 85.6 54.0 81.9 59.1

Neural

Net

57.5 59.9 85.6 54.0 71.3 67.9

Partition 57.5 59.9 85.6 54.0 91.3 68.6

Random

Forest

57.5 60.0 85.6 54.0 81.9 54.0

7. Discussion and Conclusion

In this paper, we have conducted correlation and classification

analyses for a dataset containing heart disease data

for 303 patients. We find that, among the numerical variables

in the dataset, all pairwise correlations are weak, the

strongest existing between age and maximum heart rate.

These variables exhibit a weak negative correlation, which

could exist for multiple reasons. For example, it has been

found that aging slows the natural electrical activity of the

heart’s pacemaker, the sinoatrial (SA) node [11]. The cause

has not been exactly identified, but it is suspected to be due

to changes in the ion channels of sinoatrial myocytes, cells

in the SA node. And yet, the correlation is not at all strong.

This is an indication that multiple factors contribute to

maximum heart rate and similar measures of heart health,

including exercise and diet. Without knowing the exact

details of the patients whose data were used, we suspect

that the considerable scatter in the maximum heart rateage

regression is due to a variety of lifestyles. This claim

is supported by the relatively wide range taken on by the

histograms in Figure 3.

By creating linear regressions with ST segment depres-

MATHEMATICS AND COMPUTER SCIENCE

Broad Street Scientific | 2019-2020 | 67


sion as a dependent variable, we have found that weak

positive correlations exist between it and blood cholesterol

levels, age, and resting blood pressure, while a negative

correlation exists between it and maximum heart rate. We

can theorize why these correlations exist, but more data

would be required to confirm their statistical significance.

High blood cholesterol levels are a known risk factor for

atherosclerosis and CAD due to the formation of fatty

plaques blocking blood vessels. Given that ST depression

is an indication of these diseases, the correlation is expected.

Age and high blood pressure (hypertension) are also

related to the occurrence of heart disease, though they are

not dominant risk factors. The correlation with maximum

heart rate is likely not statistically significant, as it has actually

been found that accelerated heart rate (e.g. tachycardia)

is a risk factor for heart disease, particularly in men [12].

We have also demonstrated the ability to apply classification

machine learning algorithms to predict the presence

or absence of heart disease. Using Gini coefficients and

jitterplots, we are able to ascertain the variables most important

in the classification. Here, the most important numerical

variable was, as expected, ST segment depression,

the numerical variable that gave the most comprehensive

indication of cardiovascular disease. Using this variable,

it was found that the partition model performed the best,

predicting absence of heart disease with an accuracy of

over 90% and presence with an accuracy of about 70%.

The primary limitations of this work include the lack

of sufficient data to draw statistically significant conclusions

and the inability to attribute heart disease to a limited

number of causes or risk factors. We have seen the high

level of variance in this dataset, and a principal component

analysis (PCA) may provide a way to dimensionally reduce

the data to only a few key variables. In addition, gathering

more data on variables including, but not limited to,

genetics and race may provide a more holistic and accurate

model to predict cardiovascular disease. Nevertheless,

this research takes the first step in expanding classification

machine learning techniques to a clinical setting, and we

envision it helping automate the process of heart disease

diagnosis.

8. Acknowledgements

The author would like to acknowledge the help and support

of Mr. Robert Gotwals of the North Carolina School

of Science and Mathematics (NCSSM) and the NCSSM

Online Program for making this research project possible.

Research (2018). https://www.mayoclinic.org/diseases-conditions/angina/symptoms-causes/syc-20369373

[3] Publishing, H.H.: The heart attack gender gap. https://

www.health.harvard.edu/heart-health/the-heart-attackgender-gap

[4] Cholesterol and Heart Disease. WebMD (2018).

https://www.webmd.com/heart-disease/guide/heart-disease-lower-cholesterol-risk#1

[5] Finkelhor, R.S., Newhouse, K.E., Vrobel, T.R., Miron,

S.D., Bahler, R.C.: The st segment/heart rate slope as a predictor

of coronary artery disease: comparison with quantitative

thallium imaging and conventional st segment criteria.

American heart journal 112(2), 296–304 (1986)

[6] McSharry, P.E., Clifford, G.D., Tarassenko, L., Smith,

L.A.: A dynamical model for generating synthetic electrocardiogram

signals. IEEE transactions on biomedical engineering

50(3), 289–294 (2003)

[7] KNN Classification using Scikit-learn. https://www.

datacamp.com/community/tutorials/k-nearest-neighbor-classification-scikit-learn

[8] Yiu, T.: Understanding Random Forest. Towards Data

Science (2019). https://towardsdatascience.com/understanding-random-forest-58381e0602d2

[9] Ray, S., Analytics, B.: 6 Easy Steps to Learn Naive Bayes

Algorithm (with code in Python) (2019). https://www.analyticsvidhya.com/blog/2017/09/naive-bayes-explained/

[10] Gini Index For Decision Trees. QuantInsti (2019).

https://blog.quantinsti.com/gini-index/

[11] Larson, E.D., Clair, J.R.S., Sumner, W.A., Bannister,

R.A., Proenza, C.: Depressed pacemaker activity of sinoatrial

node myocytes contributes to the age-dependent

decline in maximum heart rate. Proceedings of the National

Academy of Sciences 110(44), 18011–18016 (2013)

[12]Perret-Guillaume, C., Joly, L., Benetos, A.: Heart rate

as a risk factor for cardiovascular disease. Progress in cardiovascular

diseases 52(1), 6–10 (2009)

9. References

[1] Angina Pectoris (2019). https://emedicine.medscape.

com/article/150215-overview

[2] Angina. Mayo Foundation for Medical Education and

68 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


MODELING THE EFFECT OF STEM CELL-TARGETING

IMMUNOTHERAPY ON TUMOR SIZE

Amber Pospistle

Abstract

Cancer stem cells (CSCs) are associated with aggressive tumors and are believed to be a driving factor in tumor growth

due to their ability to differentiate and their high reproductivity [17]. Immunotherapy treatments targeting CSCs have

been shown to be promising in reducing tumor size in experimental studies involving mice [15]. In this paper, a computational

model of stem cell-targeted immunotherapy is proposed. The interaction of cytotoxic T-cells (CTCs) and dendritic

cells (DCs) with cancer cells is modeled. The ordinary differential equation model will be used to assess the effectiveness of

dendritic cell vaccines and T-cell adoptive therapy with or without chemotherapy in reducing tumor size and growth. The

results of the model show that immunotherapy treatments combined with chemotherapy are the most effective treatment

for reducing tumor size and growth. The model confirms that CSCs are likely a driving factor in tumor growth.

1. Introduction

Immunotherapy is one of the newest treatments for

cancers. According to the World Health Organization,

the number of new cancer cases is expected to increase by

70 percent over the next 20 years [10]. Immunotherapy

is effective in treating several types of cancer including

non-Hodgkin’s lymphoma, multiple myeloma, prostate

cancer, renal cell carcinoma, malignant melanoma, colorectal

cancer, and small-cell lung cancer [13]. This treatment

stimulates the adaptive or acquired immune system

to kill tumor cells and has fewer side effects than chemotherapy

[19]. For patients with melanoma, a type of cancer

unresponsive to chemotherapy, immunotherapy may be

more effective [19]. Immunotherapy is most effective for

patients in the early stages of cancer [19]. There are multiple

types of immunotherapy including dendritic cell vaccines,

adoptive T-cell therapies, and stimulating the immune

system through vaccines and/or cytokines [17] [19].

Tumor cells have specific antigens that trigger immune

responses [1]. Tumor antigens are often derived from viral

proteins, point mutations, or encoded by cancer-germline

genes. Antigens may be the product of oncogenes or tumor

suppressor genes, over-expressed genes, products of oncogenic

viruses, or oncofectal antigens. Tumor antigens may

be present on the cell surface or in the bloodstream. Most

types of cancer have identified antigens. Antigens are classified

as tumor-specific (only associated with tumor cells)

or tumor-associated (associated with both tumor cells and

normal cells) [20]. Different types of cancer and different

patients have different tumor antigens. When antigens are

recognized by lymphocytes (white blood cells), T lymphocytes

(T-cells) multiply and kill the cancer cells [10].

According to the Tumor Immune Surveillance hypothesis,

immune cells including monocytes, macrophages,

dendritic cells (DCs), and natural killer cells provide a

short-lived response that kills tumor cells and captures debris

from dead tumor cells. T-cells and B cells (B lymphocytes)

provide long-lived antigen-specific responses and

have an effective memory [3] [10]. Dendritic cells allow

T-cells to recognize an antigen. The effectiveness of dendritic

cells in presenting an antigen to T-cells influences

the effectiveness of immunotherapy. As a result, dendritic

cells are a major target when developing immunotherapy

treatments [19].

In addition to T-cell receptors recognizing an antigen,

the T-cell needs to receive a costimulatory signal to become

activated. When a T-cell is activated, T-cell receptors

bind to antigen peptides on major histocompatibility

complex class (MHC) molecules presented by dendritic

cells or macrophages [10] [17]. In order for the T-cell receptor

to bind to the class I MHC molecules, a glycoprotein

called CD8 is needed. Once the T-cell receptors bind to

the cell with the antigen, the cytotoxic T-cell will release

cytokines (chemical messengers) that have antitumor effects.

The cytotoxic T-cell will also release cytotoxic granules

including perforin and granzymes. Perforin forms a

pore in the cell membrane and allows granzymes to enter

the cell which leads to cell apoptosis. Cytotoxic T-cells can

also kill the target cell when FasL (Fas ligand) on the T-cell

surface binds with Fas on the target cell surface. Activated

CD4+ T- cells (also known as T-helper cells) can secrete

cytokines such as interleukins to promote the growth of

the cytotoxic T-cell population [10][17].

Initially, these immune cells can destroy tumor cells

or entire tumors. However, pathways such as PD-L1 and

PD-L2 can inhibit the activation of T-cells [17]. PD-L1 is

expressed on the surface of up to 30 percent of solid tumors

[14]. Tumor cells develop mechanisms to evade

immunosurveillance including producing immuno suppresive

cytokines or altering their expression of interleukins

which may cause inactivation or prevent maturity of

DCs. In addition, CD4+ and CD8+ T-cell responses may

be suppressed by the immune system to prevent damage to

MATHEMATICS AND COMPUTER SCIENCE

Broad Street Scientific | 2019-2020 | 69


healthy cells. Regulatory cells including regulatory T-cells,

proteins, and natural suppressor cells can suppress immune

system response [21]. Regulatory T-cells play a negative

regulation role on immune cells including cytotoxic T-cell

lymphocytes [19]. T helper type 2 cells, neutrophils (a type

of white blood cell), and activated M2 macrophages can

inhibit cytotoxic T-cell immune response [16]. Therefore,

regulatory T-cells play a major role in the effectiveness of

immunotherapy treatments.

Tumor cells may change their expression of antigens

to evade immune response [3]. Immunotherapy blocks

inhibitory pathways used by tumors [17], in contrast to

targeted kinase inhibitors which are prone to becoming

unresponsive after tumor cells evade it. Immunotherapy is

more effective over time as it activates the body’s immune

system and does not target one specific characteristic of a

constantly evolving tumor [14].

Multiple types of immunotherapy have been studied

in experimental and clinical trials. Immunotherapy treatments

are often given in combination for increased effectiveness.

Dendritic cell vaccines are one promising treatment.

To develop a dendritic cell vaccine, dendritic cells

are extracted from the patient. The cells from the patient

are then injected with tumor antigens and inserted back

into the patient in the form of a vaccine. The immune response

against the vaccine causes the immune system to

recognize the tumor antigen and kill the tumor cells [2]

[17]. Without the vaccine, the patient’s immune system

would not develop a strong response against the tumor

antigen, as the antigen belongs to the patient. Dendritic

cells are the best type of cell for presenting tumor antigens

to other cells such as effector cells or T-cells [2].

A similar treatment with T-cells, called T-cell adoptive

therapy, has also become a promising immunotherapy

treatment. In adoptive T-cell therapy, naive T-cells from

the patient are isolated. The naive T-cells are activated in

vitro by DCs with the tumor antigen. The activated T-cells

with the antigen are readministered to the patient [17].

Most adoptive therapies use CD8+ T-cells that destroy a

tumor cell with a specific antigen by binding to its complex

of class I major histocompatibility (MHC) proteins

[10]. The goal of T-cell adoptive therapy is to increase the

activated T-cell population with receptors that recognize

antigens specific to the tumor. Previous research has found

that T-cell adoptive therapy can destroy large tumors [17]

[21] [23]. In this paper, T-cell adoptive therapy will be referred

to as T-cell treatment.

Many solid tumors are heterogeneous, containing both

cancer stem cells (CSCs) and non-CSCs (nCSCs). CSCs can

self-renew and differentiate into or produce other types of

cancer cells [17]. CSCs have been implicated as a driving

factor in tumor growth and recurrence of tumors [4] [17].

CSCs have been identified for many types of cancer including

head and neck squamous cell carcinoma, pancreatic,

breast, and lung cancer [17]. However, cancer treatments

including chemotherapy and radiation have been ineffective

in killing CSCs. As a result, cancer stem-cell based

immunotherapies are of increasing interest in cancer research.

Due to the complexity of the tumor microenvironment

and the high cost of cancer research, computational models

of the cancer immune system have been created to model

the effectiveness of immunotherapy compared to traditional

treatments [17]. A past computational model of

dendritic cell vaccinations found that the treatment is most

effective when given multiple times based on a set interval.

Two computational models were proposed to find the optimal

time for dendritic cell vaccinations [2][19]. Multiple

models of the interaction of T-cells and cancer [8][9] and

other mathematical models of tumor-immune interaction

[1][5][6][7] have been proposed. However, there are few

computational models of T-cell adoptive therapy or stem

cell-targeted immunotherapy.

This paper will model the interaction of cancer stem

cells, non-cancer stem cells, cytotoxic T-cells (CTCs), and

dendritic cells (DCs) through ordinary differential equations.

This work will model the effect of CSCs and nCSCs

on tumor size, determine the cell population in which

dendritic vaccinations are the most effective, model T-cell

treatment with inoculation of different cell populations,

and determine how chemotherapy and immunotherapy

together impact tumor size.

2. Computational Approach

For this paper, the computational model was created

using STELLA Architect [18], a differential equation solver

that can model the different cell populations, inflows

(replication), and outflows (cell death). The model consists

of two cancer cell populations: CSCs and nCSCs. Populations

of DCs and CTCs specific to CSCs and nCSCs were

included in the model. A DC can be immature (doesn’t

have tumor antigen), mature with CSC-specific antigens,

or mature with nCSC-specific antigens. A CTC can be

considered naive, mature, CSC-specific, or nCSC-specific.

Several assumptions were made in the model:

1. CSCs self-renew and nCSCs can renew at a varying

rate based on the type of cancer.

2. CSCs can become two nCSCs and a nCSC can

become a CSC. A CSC can also become two CSCs

through replication or become one CSC and one

nCSC.

3. Immature DCs can become mature through the

consumption of antigens. If the immature DC

consumes a CSC or a nCSC, the mature DC will be

CSC-specific or nCSC-specific respectively.

4. Mature DCs in the model will present antigens

to naive CTCs causing the CTCs to become

activated. The CTCs will be able to kill cells with

the same antigens as the DC. The CTC will be

70 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


Table 1. List of parameters used in the ordinary differential equations [17]

Parameter Parameter Meaning Parameter Value

S Total Population of CSCs See Equation 1

P Total Population of nCSCs See Equation 2

T s

Total Population of Activated CSC-specific CTCs See Equation 3

T p Total Population of Activated nCSC-specific CTCs See Equation 4

D s

Total Population of Mature CSC-specific DCs See Equation 5

D p Total Population of Mature nCSC-specific DCs See Equation 6

C Concentration of Chemotherapy See Equation 7

α S

Growth rate of CSC population due to symmetric replication See Figure Captions

ρ PS

Transition rate from nCSCs to CSCs See Figure Captions

ρ SP

Transition rate from CSCs to nCSCs See Figure Captions

β S

Death rate of CSCs due to CSC-specific CTCs 6.2 × 10 -8 cells -1 day -1

δ S

Death rate of CSCs due to natural causes See Figure Captions

Γ S

Death rate of CSCs due to chemotherapy 1.4 × 10 -3 day -1 (µg/mL) -1

α P

Growth rate of nCSCs due to replication See Figure Captions

α SP

Growth rate of nCSC population due to asymmetric division of CSCs See Figure Captions

β P

Death rate of nCSCs due to nCSC-specific CTCs 6.2 × 10 -8 cells -1 day -1

δ P

Death rate of nCSCs due to natural causes See Figure Captions

Γ P

Death rate of nCSCs due to chemotherapy 5.0 × 10 -3 day -1 (µg/mL) -1

χ TS T n S

Saturated Activation Rate of CTCs due to mature CSC-specific DCs 4.5 × 10 4 aCTCs day -1

T s

Mature CSC-specific DC EC50 for CTC Activation Rate 2.5 × 10 4 mDCs

δ Ts

Death Rate of Activated CSC-specific CTCs due to natural causes 0.02 day -1

χ TS T n P

Saturated Activation Rate of CTCs due to mature nCSC-specific DCs 4.5 × 10 4 aCTCs day -1

T P Mature nCSC-specific DC EC50 for CTC Activation Rate 2.5 × 10 4 mDCs

δ TP Death Rate of Activated nCSC-specific CTCs due to natural causes 0.02 day -1

γ DS D Maturation rate of CSC-specific DCs due to consumption of cancer cells 0.0063 day -1 cancer cell -1

β DS Death Rate of CSC-specific DCs due to CSC-specific CTCs 6.2 × 10 -8 cells -1 day -1

δ DS Death Rate of CSC-specific DCs due to natural causes 0.2 day -1

γ DP D Maturation rate of nCSC-specific DCs due to the consumption of cancer cells 0.0063 day -1 cancer cell -1

β DP Death Rate of nCSC-specific DCs due to nCSC-specific CTCs 6.2 × 10 -8 cells -1 day -1

δD P

Death Rate of nCSC-specific DCs due to natural causes 0.2 day -1

e c

Elimination Rate of Chemotherapy 50 day -1

CSC-specific or nCSC-specific based on the antigen

of the DC cell.

5. All cells die.

In STELLA, seven stocks were used to represent the following

populations: CSCs, nCSCs, activated nCSC-specific

CTCs, activated CSC-specific CTCs, mature CSC-specific

DCs, mature nCSC-specific DCs, as well as the concentration

of chemotherapeutic agent.

In this paper, five models involving tumor inoculation,

dendritic cell vaccines, adoptive T-cell therapy, and chemotherapy

will be simulated. Each model will be discussed

individually and has a different STELLA model. The following

ordinary differential equations using variables defined

in Table 1 [17] were used:

(1)

(2)

(3)

(4)

(5)

MATHEMATICS AND COMPUTER SCIENCE

Broad Street Scientific | 2019-2020 | 71


Eq. 1 represents the growth of CSCs through symmetric

replication, the transition of nCSC to CSC, the transition

of CSCs to nCSCs, killing of CSCs by CTCs, natural

death of CSCs, and the death of CSCs by chemotherapy.

Eq. 2 represents the growth of nCSCs through asymmetric

replication, growth of the nCSC population through replication,

the division of a CSC into two nCSCs, killing of

nCSCs by nCSC-specific CTCs, natural death of nCSCs,

and chemotherapy killing nCSCs. Eq. 3 represents the activation

of naive CTCs (which become saturated) by mature

CSC-specific DCs and the natural death of activated

CSC-specific CTCs. Eq. 4 represents the activation of naive

CTCs (which become saturated) by mature nCSC-specific

DCs and the natural death of activated nCSC-specific

CTCs. Eq. 5 represents the maturation of immature DCs

when they encounter tumor cells with CSC-specific antigens

and the killing of mature CSC-specific DCs. Eq. 6

represents the maturation of immature DCs when they encounter

tumor cells with nCSC-specific antigens and the

killing of mature nCSC-specific DCs. Eq. 7 represents the

elimination of chemotherapy from the body. Host- specific

parameters will be in the figure captions in the results section.

These host-specific parameters vary based on patient

and type of cancer.

3. Model 1: Tumor Inoculation with CSCs and nCSCs

In Model 1, the effect of tumor inoculation with CSCs

and nCSCs on tumor size was studied. The tumor was inoculated

with 50,000 CSCs or 50,000 nCSCs on Day 0. All

cell populations were set to 0 in STELLA except for the cell

population that was inoculated. The simulation in STEL-

LA started on Day 12 (the time at which cancer cells would

take root in a host) and ended on Day 25.

4. Model 2: Dendritic Cell Vaccine Before Tumor Inoculation

In Model 2, the effects of a dendritic cell vaccine prior to

tumor inoculation on tumor size were modeled. Dendritic

cell vaccines were given on Day 0, Day 7, and Day 14. The

tumor was inoculated on Day 22. There were four types

of treatments simulated: no DC vaccination, DC vaccination

with CSC-specific DCs and nCSC-specific DCs, DC

vaccination with CSC-specific DCs, and DC vaccination

with nCSC-specific DCs. In all simulations, a population

of 10,000 CSCs and 90,000 nCSCs were used to inoculate

the tumor. On each day the vaccine was given, 1 million

mature DCs were added to the appropriate DC population

for the DC vaccination with CSCs and the vaccination

with nCSCs. For the mixed DC vaccine treatment, each

(6)

(7)

vaccine contained 100,000 CSC- specific DCs and 900,000

nCSC-specific DCs. For the no treatment simulation, the

tumor was only inoculated. The DC vaccines were simulated

in STELLA using the PULSE option. Prior to inoculation,

all cell populations except for the appropriate

mature DC population, if applicable, were set to zero. The

inoculation of the tumor was modeled in STELLA with

PULSE. The simulation was run from Day 0 to Day 49.

5. Model 3: T-Cell Treatment After Tumor Inoculation

In Model 3, the effects of T-Cell treatment after tumor

inoculation on tumor size were modeled. T-Cell treatment

was given on Day 1, Day 8, and Day 15. There were four

types of treatments: no treatment, mixed treatment with

both CSC-specific and nCSC-specific CTCs, treatment

with CSC-specific CTCs, and treatment with nCSC-specific

CTCs. T-cell treatments consisted of 1,000,000 CSCs

or nCSCs. For mixed treatment, treatments consisted of

50,000 CSC-specific CTCs and 950,000 nCSC-specific

CTCs. All cell populations were set to 0 for Day 0. T-Cell

treatments were modeled in STELLA using PULSE. The

inoculation of the tumor on Day 8 was modeled in STEL-

LA using PULSE which allowed for an inflow of 5,000

CSCs and 95,000 nCSCs.

6. Model 4: Modeling the Effectiveness of T- Cell Treatment

and Chemotherapy

In Model 4, there were six treatments modeled: no treatment,

chemotherapy only, CSC-specific T-Cell treatment

only, nCSC-specific T-cell treatment only, combined chemotherapy

and CSC-specific T-cell treatment, and combined

chemotherapy and nCSC-specific T-cell treatment.

The tumor was inoculated on Day 0 with 5,000 CSCs and

95,000 nCSCs. All other populations were set to zero. For

nCSC-specific T-cell treatment and CSC-specific T-cell

treatment, one million activated CTCs were added to the

appropriate population on Day 20 and 27 using PULSE.

For chemotherapy treatment, one injection of 6,000 µg/

mL chemotherapeutic agent was given per day from days

20-24 and 27-31. Chemotherapy was modeled in STEL-

LA using PULSE. For combined chemotherapy and T-cell

treatment, the methods for modeling chemotherapy and

T-cell treatment were used together in the STELLA model.

7. Model 5: Modeling the Effectiveness of DC Vaccines

and Chemotherapy

In Model 5, there were six treatments modeled: no

treatment, chemotherapy only, mature CSC-specific DC

vaccines only, nCSC-specific DC cell vaccines only, combined

chemotherapy and CSC-specific DC vaccines, and

combined chemotherapy and nCSC-specific DC vaccines.

72 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


The tumor was inoculated on Day 0 with 5,000 CSCs and

95,000 nCSCs. All other populations were set to zero. For

nCSC-specific DC vaccines and CSC-specific DC vaccines,

one million mature DCs were added to the appropriate

population on Day 20 and 27 using PULSE. For chemotherapy

treatment, one injection of 6,000 µg/mL chemotherapeutic

agent was given per day from days 20-24 and

27-31. Chemotherapy was modeled in STELLA using

PULSE. For combined chemotherapy and DC vaccines,

the methods for modeling chemotherapy and DC vaccines

were used together in the STELLA model.

8. Results and Discussion

For all models, the effectiveness of the treatment is

shown in terms of tumor size. Tumor size was calculated

by adding together the CSC population and nCSC population

based on the results of the STELLA model and dividing

by 100,000 using Excel [11]. All plots were created using

Mathematica [22]. The host-specific parameters used

for each model are reported in the figure captions. Refer

to Table 1 for the meaning of each parameter and its value.

both CSCs or nCSCs or only nCSCs. All T-cell treatments

were effective in reducing tumor size when compared to

no treatment.

Figure 2. This figure shows the evolution of tumor

size after dendritic cell vaccines with CSC-specific

DCs and/or nCSC-specific DCs. The host-specific parameters

are α S = 0.6 day-1 , δ S

= 0.2 day -1 , ρ SP

= 0.1 day -1 ,

α P

= 0.2 day -1 , δ P

= 0.14 day -1 , α SP

= 2.0 day -1 , and ρ PS

= 4.4

× 10 −4 day -1 .

Figure 1. This figure shows the evolution of tumor

size after inoculation with 50,000 CSCs or nCSCs. The

host-specific parameters are α S

= 0.5 day -1 , δ S

= 0.2 day -1 ,

ρ SP

= 0.15 day -1 , α P

= 0.2 day -1 , δ P

= 0.15 day -1 , α SP

= 1.8 day -1 ,

and ρ PS

= 5.3 × 10 −4 day -1 .

Inoculating a tumor with cancer stem cells leads to

significant tumor growth, and inoculating a tumor with

nCSCs leads to a significantly lower and comparatively

negligible tumor growth in the same host-specific conditions

(Fig. 1). These data support the finding that CSCs

drive tumor growth [17].

DC vaccines prior to tumor inoculation are effective in

slowing tumor growth when compared to no treatment

(Fig. 2). The CSC-specific DC vaccine was most effective

in reducing tumor size compared to the mixed DC vaccine

and nCSC DC vaccine. The mixed DC vaccine was the second

most effective in reducing tumor size.

CSC T-Cell treatment is significantly more effective in

reducing tumor size compared to the mixed T-cell treatment,

nCSC T-cell treatment, and no treatment (Fig. 3).

In other words, activated CTCs cultured with mature dendritic

cells that were pulsed with CSCs were more effective

in reducing tumor growth compared to those pulsed with

MATHEMATICS AND COMPUTER SCIENCE

Figure 3. This figure shows the evolution of tumor

size after different T-cell treatments and no treatment.

The host-specific parameters are α S

= 0.7, δ S

=

0.19 day -1 , ρ SP

= 0.24 day -1 , α P

= 0.2 day -1 , δ P

= 0.1 day -1 , α SP

=

2.8 day -1 , and ρ PS

= 1.3 × 10 −4 day -1 .

Figure 4. This figure shows the evolution of tumor

size after T-cell treatments, chemotherapy, and

combined T-cell and chemotherapy treatments. The

host-specific parameters are α S

= 0.5 day -1 , δ S

= 0.2 day -

1

, ρ SP

= 0.15 day -1 , α P

= 0.2 day -1 , δ P

= 0.15 day -1 , α SP

= 1.8 day-

1

, and ρ PS

= 5.3 × 10 −4 day -1 .

Chemotherapy is a more effective treatment for treating

tumors compared to only CSC T-cell treatment or

nCSC T-cell treatment (Fig. 4). The combined chemotherapy

treatments and T-cell treatments were most effective

in reducing tumor size. CSC T-cell treatment with

chemotherapy was the most effective followed by com-

Broad Street Scientific | 2019-2020 | 73


bined nCSC T-cell treatment with chemotherapy and chemotherapy

only. All treatments had the same tumor size

until Day 20, when the first chemotherapy and/or immunotherapy

treatment was given. All treatments involving

chemotherapy saw an oscillating decrease in tumor size

before an increase in tumor size after the last day of chemotherapy

treatment. All treatments with chemotherapy

had similar tumor size at the end of the simulation which

was significantly less than treatments without chemotherapy.

Treatments that did not involve chemotherapy saw a

significant increase in tumor size, which continued until

the end of the simulation.

Figure 5. This figure shows the evolution of tumor

size after DC vaccines, chemotherapy, and combined

chemotherapy treatment and DC vaccines. The

host-specific parameters are α S

= 0.5 day -1 , δ S

= 0.2 day -1 ,

ρ SP

= 0.15 day -1 , α P

= 0.2 day -1 , δ P

= 0.15 day -1 , α SP

= 1.8 day -1 ,

and ρ PS = 5.3 × 10−4 day -1 .

Treatments with chemotherapy, especially those combined

with immunotherapy treatments, are more effective

in reducing tumor size compared to no treatment or only

immunotherapy treatment (Fig. 4 and 5). The combined

treatment with CSC-specific DC Vaccine treatment and

chemotherapy was most effective in reducing tumor size

(Fig. 5). The combined treatment with nCSC-specific DC

vaccine treatment and chemotherapy was the second most

effective treatment for reducing tumor size. However, all

chemotherapy treatments had a very similar tumor size at

the end of the simulation. In addition, the immunotherapy

treatments and no treatment had a very similar tumor size

by the end of the simulation.

All results suggest that immunotherapy treatments

targeting CSCs are the most effective in reducing tumor

growth. After CSC-specific immunotherapy treatment,

T-cell receptors will recognize CSC-specific antigens, bind

to antigen peptides on MHC Class I molecules on CSCs,

and kill the CSCs through the release of cytokines, cytotoxic

granules, or the FasL ligand. CSC-specific dendritic

cell vaccines and T-cell adoptive therapy allow the immune

system to directly target and kill CSCs. In contrast,

nCSC-specific dendritic cell vaccines and T-cell adoptive

therapy are less effective in reducing tumor size, since

cytotoxic T-cells will kill nCSCs and lead to comparably

negligible tumor growth due to lower renewal rate and inability

to differentiate [17] (Fig. 1).

9. Conclusions

In this paper, a computational model was used to

study how dendritic cells and cytotoxic T-cells impact

tumor-immune interaction with CSCs and nCSCs. The

model found that tumor inoculation with CSCs produced

larger tumors compared to tumors inoculated with nCSCs.

CSC-specific DC vaccines are more effective in reducing

tumor growth than mixed or nCSC-specific DC vaccines

(Fig. 2). CSC-specific T-cell treatment was more effective

in reducing tumor growth than mixed or nCSC-specific

T-cell treatment (Fig. 3). Cancer treatments involving

chemotherapy were most effective in reducing tumor

growth compared to immunotherapy treatments alone

(Fig. 4 & 5). However, combined chemotherapy and immunotherapy

treatments (CSC-specific DC vaccines or

CSC-specific T-cell treatment) were most effective in reducing

tumor size. T-cell treatment was more effective in

reducing tumor size after inoculation compared to dendritic

cell vaccines. The CSC-specific T-cell treatment and

chemotherapy treatment was the most effective treatment

overall for reducing tumor growth after inoculation. This

model could be improved to include the role of T-helper

cells in activating cytotoxic T-cells [19] [17]. Mature

den- dritic cells present antigens to both cytoxic T-cells

and T-helper cells through MHC Class I and MHC Class

II [17]. T-helper cells give cytotoxic T-cells a costimulatory

signal that is needed for T-cells to become activated

as well as proliferate [17]. The model could add the role of

T-regulatory cells that inhibit T-cell activation and proliferation

[16][17][19]. In the model, exponential growth

was assumed that may not occur in vivo. Some cancer cells

that were inoculated may have been destroyed in vivo prior

to the cancer cells taking root which may be a potential

source of error. In addition, DC cells can lead to death of

CSC and nCSCs but was not considered in the model, since

the effect on the cell populations is considered negligible

compared to cell death caused by CTCs. Tumors are also

believed to affect the maturation of dendritic cells and cytotoxic

T-cells, which was not considered in the model as

the effects are currently unclear. The proposed computational

model could have large implications in studying immunotherapy

effectiveness and combined immunotherapy

and chemotherapy treatments for many types of cancer

including pancreatic cancer, breast cancer, squamous head

and neck cancer, and brain tumors. The host-specific parameters

and chemotherapy treatment parameters can be

modified for a patient or a type of cancer. In addition, the

renewal rate of CSCs and nCSCs can be modified for each

type of cancer.

10. Acknowledgements

The author thanks Mr. Robert Gotwals for his assistance

with this work and the Broad Street Scientific editors.

74 | 2019-2020 | Broad Street Scientific MATHEMATICS AND COMPUTER SCIENCE


11. References

[1] Al-Tameemi, M., Chaplain, M., d’Onofrio, A. (2012).

Evasion of tumours from the control of the immune system:

consequences of brief encounters. Biology direct,

7(1), 31.

[2] Castiglione, F., Piccoli, B. (2006). Optimal control in a

model of dendritic cell transfection cancer immunotherapy.

Bulletin of Mathematical Biology, 68(2), 255-274.

[3] Cheng, M., Chen, Y., Xiao, W., Sun, R., Tian, Z.

(2013). NK cell-based immunotherapy for malignant diseases.

Cellular molec- ular immunology, 10(3), 230.

[4] Dashti, A., Ebrahimi, M., Hadjati, J., Memarnejadian,

A., Moazzeni, S. M. (2016). Dendritic cell based immunotherapy

using tumor stem cells mediates potent antitumor

immune responses. Cancer letters, 374(1), 175-185.

[5] DePillis, L. G., Eladdadi, A., Radunskaya, A. E. (2014).

Modeling cancer-immune responses to therapy. Journal of

pharmacokinetics and pharmacodynamics, 41(5), 461-478.

[6] de Pillis, L. G., Radunskaya, A. (2003). A mathematical

model of immune response to tumor invasion. In Computational

Fluid and Solid Mechanics 2003 (pp. 1661-1668).

Elsevier Science Ltd.

[7] de Pillis, L. G., Radunskaya, A. E., Wiseman, C. L.

(2005). A validated mathematical model of cell-mediated

immune response to tumor growth. Cancer research,

65(17), 7950-7958.

[8] Kirschner, D., Panetta, J. C. (1998). Modeling immunotherapy

of the tumor–immune interaction. Journal of

mathematical biology, 37(3), 235-252.

[9] Kuznetsov, V. A., Makalkin, I. A., Taylor, M. A., Perelson,

A. S. (1994). Nonlinear dynamics of immunogenic

tumors: parameter estimation and global bifurcation analysis.

Bulletin of mathematical biology, 56(2), 295-321.

[10] Mahasa, K. J., Ouifki, R., Eladdadi, A., de Pillis, L.

(2016). Mathematical model of tumor–immune surveillance.

Journal of theoretical biology, 404, 312-330.

[11] Microsoft Excel [Software] Available from https://

products.office.com/en-us/excel.

[12] de Mingo Pulido, A., Ruffell, B. (2016). Immune regulation

of the metastatic process: implications for therapy.

In Advances in cancer research (Vol. 132, pp. 139-163).

Academic Press.

[13] Morisaki, T., Matsumoto, K., Onishi, H., Kuroki,

H., Baba, E., Tasaki, A., Tanaka, M. (2003). Dendritic

cell-based combined immunotherapy with autologous tumor-pulsed

dendritic cell vaccine and activated T cells for

cancer patients: rationale, current progress, and perspectives.

Human cell, 16(4), 175-182.

[14] Ohaegbulam, K. C., Assal, A., Lazar- Molnar, E., Yao,

Y., Zang, X. (2015). Human cancer immunotherapy with

antibodies to the PD-1 and PD-L1 pathway. Trends in

molecular medicine, 21(1), 24-33.

[15] Palucka, K., Banchereau, J. (2012). Cancer immunotherapy

via dendritic cells. Nature Reviews Cancer, 12(4),

265.

[16] Rhodes, A., Hillen, T. (2019). A Mathematical Model

for the Immune-Mediated Theory of Metastasis. bioRxiv,

565531.

[17] Sigal, D., Przedborski, M., Sivaloganathan, D., Kohandel,

M. (2019). Mathematical modelling of cancer stem

cell-targeted immunotherapy. Mathematical Biosciences,

318, 108269.

[18] STELLA Architect (Version 1.6.2) [Software] Available

from https://www.iseesystems.com/store/products/

stellaarchitect.aspx.

[19] Qomlaqi, M., Bahrami, F., Ajami, M., Hajati, J. (2017).

An extended mathematical model of tumor growth and its

interaction with the immune system, to be used for developing

an optimized immunotherapy treatment protocol.

Mathematical biosciences, 292, 1-9.

[20] Vigneron, N. (2015). Human tumor antigens and cancer

immunotherapy. BioMed research international, 2015.

[21] Wilson, S., Levy, D. (2012). A mathematical model

of the enhancement of tumor vaccine efficacy by immunotherapy.

Bulletin of mathematical biology, 74(7), 1485-

1500.

[22] Wolfram Mathematica (12.0 Student Version) [Software]

Available from https://www.wolfram.com/mathematica/.

[23] Yu, J. L., Jang, S. R. J. (2019). A mathematical model

of tumor-immune interactions with an immune checkpoint

inhibitor. Applied Mathematics and Computation,

362, 124523.

MATHEMATICS AND COMPUTER SCIENCE

Broad Street Scientific | 2019-2020 | 75


THE COSMIC RADIATION SHIELDING PROPERTIES OF

LUNAR REGOLITH

Eleanor Murray

Abstract

Lunar regolith is the most accessible material for use as radiation shielding for human habitation on the Moon. This study

aims to determine the thickness of lunar regolith shielding necessary to protect humans from cosmic radiation and its secondaries,

including neutrons. We measured the percentage of thermal neutrons that passed through prepared LHS-1 Lunar

Highlands Simulant samples of different thicknesses using the Neutron Powder Diffraction Facility at the PULSTAR

Reactor of NC State University, prepared for a neutron transmission experiment. In addition, we modeled an analogue

of the experiment using GEANT4 software. We present the results of the neutron experiments as well as preliminary

results of the necessary thickness at which the radiation dose will drop to safe levels based on GEANT4 simulations of the

interactions of galactic cosmic rays with the lunar regolith-like material.

1. Introduction

As NASA prepares for a Moon landing in 2024 with the

Artemis program, leading towards “sustainable missions

by 2028” [1], the In-Situ Resource Utilization (ISRU) of

lunar regolith will be integral. Particularly on a large scale,

the costs of ISRU will need to be balanced with the potential

costs of transporting materials from Earth. One potential

cost is that of launching radiation shielding materials

to the Moon. However, this cost could be mitigated by using

a material readily available on the Moon as radiation

shielding: the lunar regolith. Eckart found that a 1 meter

layer of lunar regolith would prevent crew members from

receiving more than about 3 cSV of radiation from each

solar particle event [2]. The main types of radiation on the

Moon are galactic cosmic rays (GCR) and solar particle

events (SPE). Galactic cosmic rays are a constant source of

radiation, while solar particle events occur only occasionally

but expose astronauts to a greater momentary amount

of radiation. During an SPE, astronauts can stay in an area

with more shielding, but GCR is harder to protect against

since it is constant [3]. The protons in the SPE are also

“fairly easy to stop, compared to GCR” [4]. More research

is necessary in order to protect astronauts on long-term

missions from the GCR.

2. Lunar Regolith as Radiation Shielding

On Earth, humans are protected from excessive amounts

of solar radiation by Earth’s magnetic field. The Moon’s

magnetic field is much weaker [5], so radiation shielding

will be necessary for human habitation. Eckart [2] writes

that calculations have found layers of lunar regolith 50-100

cm deep to be sufficient in decreasing radiation doses from

the GCR and SPE to acceptable levels [6]. Miller [6] found

that less than 50 cm of lunar regolith should be enough to

stop the most damaging and common galactic cosmic rays

(GCR) as well as protons from solar particle events (SPE).

This study found that lunar regolith, and several simulants

are comparable to aluminum in their radiation-shielding

properties by measuring the dose reduction per unit areal

density [6]. However, this study did not investigate the

impact of neutrons in this measurement. As galactic cosmic

rays impact the lunar surface, they produce secondary

radiation, which includes neutrons. Both the protons from

above and the neutrons from below will pose major risks

to astronauts [7]. A greater understanding of the radiation

shielding properties of lunar regolith is essential to planning

potential lunar habitats that will attenuate radiation

to safe levels.

3. Safe Limits of Radiation

Determing a safe level of radiation is difficult as studies

use varying units. A gray is a measure of how much energy

was absorbed; one gray is equal to one joule per kilogram.

Sieverts are similar but relate to the expected amount of biological

damage; the conversion between grays and sieverts

depends on the type and energy of the radiation. Globus

and Strout surveyed the literature and found that in the

context of orbital space settlements, the general population

should not receive more than 20 millisieverts per year [4].

The World Nuclear Association states that less than 100

millisieverts per year is harmless [8]. In addition, NASA

lists 500 milligray-equivalents as the limit of radiation to

the blood-forming organs of astronauts per year[9]. The

radiation on the moon comes from Solar Particle Events

(SPE), which are occasional but deliver large amounts of

radiation, and Galactic Cosmic Rays (GCR), which are

constant, deliver smaller amounts of radiation, come from

far away galaxies as opposed to the Sun, and are harder

to shield against than SPE. When these charged particles

impact the regolith of the Moon, nuclear reactions occur

and produce free neutrons, adding to the radiation dose

humans would receive [7].

76 | 2019-2020 | Broad Street Scientific PHYSICS


4. Direct Measurements of Lunar Radiation

Previous research, such as the Lunar Exploration Neutron

Detector (LEND), took data from approximately 50

kilometers above the moon instead of at its surface [10].

The LEND examined albedo neutrons, or the neutrons

that are created in the lunar surface and then bounce back

towards space [11], instead of the neutrons that are created

in the regolith and move downwards where they would

contribute to the radiation dose experienced by astronauts.

During the Apollo 17 mission, astronauts successfully

deployed the Lunar Neutron Probe Experiment (LNPE).

This probe consisted of a 2-meter-long rod that was placed

in the lunar regolith to measure the rates of capture of low

energy neutrons at various depths. This experiment is one

of the few measurements of neutrons on the lunar surface.

However, the experiment was designed with low energy

neutrons in mind, while the radiation dose from neutrons

also includes high energy neutrons. The LNPE report also

included only preliminary data; and calibration of the data

had not been completed [12].

These limitations on the available data require GEANT4

simulations as well as an empirical experiment to model

the radiation dose on the Moon as it varies with depth of

lunar regolith.

6. Method 1: Nuclear Reactor

6.1 - Methodology

We investigated the feasibility of using lunar regolith

as shielding against neutrons from the radiation environment

of the Moon. LHS-1 Lunar Highlands Simulant was

used as a proxy for lunar regolith. The simulant was contained

by aluminum boxes (show schematically in Fig. 1)

that were manufactured on the NC School of Science and

Mathematics campus.

Figure 1. Diagram of Aluminum Boxes for Samples

Ten samples that are two inches by two inches across

with varying thicknesses x were prepared. The smallest

thickness was ¼ inch, and the largest thickness was 2 ½

inches, increasing in ¼ inch increments.

5. Lunar Regolith Simulants

Lunar regolith is difficult to obtain, so most research

on lunar regolith involves JSC-1 lunar regolith simulant,

which is produced by NASA. We utilized LHS-1 Lunar

Highlands Simulant since its availability is higher than that

of JSC-1 Lunar Regolith Simulant. However, their percent

compositions are roughly similar, as shown in Table 1.

Table 1. Percent Composition by weight of oxides.

Data compiled from references [17-19].

Oxide Lunar Regolith LHS-1 JSC-1

(Highlands)

Na .6 2.30 2.7

Mg 7.5 11.22 9.01

Al 24.0 26.24 15.02

Si 45.5 44.18 47.71

Ca 15.9 11.62 10.42

Ti .6 .79 1.59

Fe 5.9 3.04 10.79

Since an estimate of the linear stopping power of a compound

can be made by assuming linear stopping power is

additive [13], the linear stopping powers of lunar regolith,

JSC-1 Lunar Regolith Simulant, and LHS-1 Lunar Highlands

Simulant should be roughly equal.

PHYSICS

Figure 2a. Image of Neutron Powder Diffraction Facility.

Image credit: Mr. Scott Lassell [14]

The samples were taken to the NC State PULSTAR reactor,

and were experimented on at the Neutron Powder

Diffraction Facility (shown in Figure 2a) prepared for

a transmission experiment. A diagram of the neutron

transmission experiment carried out on these samples

is shown in Figure 2b. The de Broglie wavelength of the

neutrons was 1.478 angstroms [15], corresponding to a kinetic

energy of approximately 0.03745 electron-volts. The

number of neutrons that hit the detector was measured

both without a target and after passing through the samples.

Broad Street Scientific | 2019-2020 | 77


The attenuation length is defined as the depth of the material

where the intensity of the radiation decreases to 1/e of

the initial intensity [16]. The attenuation length of the LHS-

1 lunar regolith simulant is approximately 6.51 g/cm 2 . The

mass attenuation coefficient is equivalent to the chance that

any one neutron will interact per centimeter divided by the

density of the material [13]. In this case, the mass attenuation

coefficient is 0.15 cm 2 /g.

Figure 2b. Diagram of Neutron Transmission Experiment

6.2 - Results

The equation (N/N 0

)=e -b*t was used to analyze the data,

where N is the number of neutrons that passed through the

sample in one minute, N 0

is the intensity of neutrons in the

initial beam in one minute, b is a constant that represents

the probability any one neutron will interact per centimeter,

and t is the thickness of the lunar regolith simulant sample

in centimeters. N=3168e (-.18*t) was found to fit the data, so for

any one neutron, there is an 18% chance that it will interact

per centimeter of lunar regolith simulant. Figure 3 shows

the thickness of lunar regolith simulant in centimeters on the

x-axis and the neutron fluence over one minute per cm 2 on

the y-axis.

7. Method 2: GEANT4 Simulations

7.1 - Methodology

GEANT4 simulation software was run in batch mode for

each thickness, and a short Python program was used to analyze

the output to count the number of neutrons that passed

through the material. Figure 5 below shows a side view of

this scenario, where the blue rectangle represents the material

defined to be similar by percent composition to LHS-1 Lunar

Highlands Simulant, the gray bar is “Shape 2,” the green tracks

show the neutrons, and the neutrons start from the middle

of the left side of the material. Only 100 neutrons are shown

for clarity. Neutrons were counted as having passed through

the material if they hit Shape 2, which was a thin cylinder

modeled as bone with a diameter of 1.27 cm, centered at the

position of the point source and thickness x centimeters from

the point source.

Figure 3. Neutron Fluence vs. Thickness of Sample

The data were also linearized to give the equation

y=-.18*t+ln(N 0

) where y=ln(N). The slope of this graph (Fig.

4) again shows that there is an 18% chance that any one neutron

will interact per centimeter of lunar regolith simulant.

Figure 4. ln(Neutron Fluence) per cm 2 vs. Thickness of

Sample

Figure 5. GEANT4 Simulation of Neutrons Passing

Through the Material.

7.2 - Results

Figure 6 shows the number of neutrons per centimeter

squared that passed through the material vs. the thickness

of the material in centimeters.

According to the equation, (N/N 0

)=e -bt , there is approximately

a 12% chance per centimeter that any particular

neutron will interact with the material. The attenuation

length is 9.8 g/cm 2 , and the mass attenuation coefficient is

0.10 cm 2 /g. Figure 7 shows the same data, but linearized,

so that y=-bt+ln(N 0

), and y=ln(N). Here, b equals approximately

0.12, so there is a 12% chance per centimeter that

a neutron will interact with the lunar regolith simulant.

78 | 2019-2020 | Broad Street Scientific PHYSICS


Figure 6. Number of Neutrons/cm 2 vs. Thickness of

Material.

discrepancy is the differences in geometry of the two experiments;

in the nuclear reactor experiment, the neutron

detector was 135 centimeters [14] away from the neutron

source, while in the GEANT4 experiment, the detector was

set right behind the sample with the neutron source set at

the front of the sample. Therefore in the nuclear reactor

experiment, the distance between the neutron source and

detector was always 135 centimeters, while in GEANT4,

the distance between the neutron source and detector

varied between 0.635 centimeters and 6.35 centimeters in

0.635 centimeter increments. Additionally, the aperture of

the collimated neutron beam at the reactor had an internal

diameter of 1.3 centimeters [14], while in GEANT4, the

neutrons came from a collimated point source. Other possible

sources of the discrepancy are the slightly different

percent compositions between the LHS-1 Lunar Highlands

Simulant used in the nuclear reactor experiment and the

material used in the GEANT4 experiment. In the future,

more simulations can be done with protons simulating the

GCR, including neutron secondaries, to find the thickness

of lunar regolith at which the radiation dose drops to safe

levels.

9. Conclusion

Figure 7. ln(Number of Neutrons) vs. Thickness of

Material.

The percent difference between the value of b, or the

percent chance that any one neutron will interact per

centimeter, found with the nuclear reactor (18%) and the

value found with GEANT4 simulations (12%) is 40%. The

percent difference in the attenuation lengths is 36%, and

the percent difference in the mass attenuation coefficients

is 40.%. A possible source of this error is the differences

in the geometry of the detectors; the neutron detector at

the nuclear reactor is set approximately 1 meter behind the

sample while the detector in GEANT4 is set directly behind

the sample.

8. Discussion

Based on the empirical method described earlier, about

6.5 g/cm 2 is the attenuation length of 0.03745 eV neutrons

in LHS-1 Lunar Highlands Simulant. For any one neutron,

there is about an 18% chance that it will interact with the

simulant per centimeter. However, a similar experiment

modeled in GEANT4 found a 12% chance per centimeter

that any one neutron will interact with the simulant, and

an attenuation length of 9.8 g/cm 2 . The percent difference

between the chances each neutron will interact per centimeter

is 40%, and the percent difference between the attenuation

lengths is 36%. A possible explanation for the

PHYSICS

Thermal neutrons have an 18% chance per centimeter

of interacting with LHS-1 Lunar Highlands Simulant, with

an attenuation length of 6.51 g/cm 2 and a mass attenuation

coefficient of .15 cm 2 /g. The radiation shielding properties

of lunar regolith should be similar to LHS-1 Lunar Highlands

Simulant since their percent compositions are similar.

This information is useful to the designs of prospective

lunar habitats. Simulations in GEANT4 show reasonable

agreement with the thermal neutron transmission experiment.

However, thermal neutrons are not a great model

for cosmic radiation, so GEANT4 will be used to model the

interactions of lunar regolith with the full radiation environment

of the Moon in order to determine the thickness

of lunar regolith at which the radiation dose falls to safe

levels for long-term human habitation.

10. Acknowledgements

The author would like to thank their mentor, Dr. Jonathan

Bennett, the physics chair at the North Carolina

School of Science and Mathematics (NCSSM), Mr. Benjamin

Wu for his assistance with coding a program for data

analysis, as well as Ms. Bec Conrad, the manager of the

Fabrication and Innovation Laboratory at NCSSM, and

the NCSSM Foundation. Additionally, the author would

like to thank Dr. Quinsheng Cai and Mr. Scott Lassell

for their assistance in utilizing the PULSTAR Reactor at

North Carolina State University, the developers and forum

of GEANT4, as well as the Center for Lunar and Asteroid

Surface Science (CLASS) Exolith Lab for providing the

Broad Street Scientific | 2019-2020 | 79


LHS-1 Lunar Highlands Simulant used in this study.

11. References

[1] “What is Artemis?” NASA.gov, https://www.nasa.

gov/what-is-artemis. Accessed 28 Oct. 2019.

[2] Eckart, Peter. The Lunar Base Handbook. The Mc-

Graw-Hill Companies, Inc., 1999

[3] Blanchett, Amy and Laurie Abadie. “Space Radiation

Won’t Stop NASA’s Human Exploration.” NASA.gov, 12

Oct. 2017, https://www.nasa.gov/feature/space-radiation-won-t-stop-nasa-s-human-exploration

[4] Globus, Al and Joe Strout. “Orbital Space Settlement

Radiation Shielding”. National Space Society, April 2017.

[5] Halekas, Jasper S. and Robert P. Lin. “Determining Lunar

Crustal Magnetic Fields and Their Origin”. Lunar and

Planetary Institute, https://www.lpi.usra.edu/meetings/

LEA/presentations/tues_pm/2_Halekas_Lunar_Crustal_

Magneti_Breakout1.pdf. Accessed 27 Oct. 2019

[13] Knoll, Glenn F. Radiation Detection and Measurement,

page 41. 3rd. ed., John Wiley and Sons, Inc., 2000

[14] Lassell, Scott. Personal interview. 23 Oct. 2019

[15] “Neutron Powder Diffraction Facility.” NCSU.

edu, https://www.ne.ncsu.edu/nrp/user-facilities/neutron-diffraction-facility/.

[16] X-Ray Attenuation Length. http://henke.lbl.gov/optical_constants/atten2.html.

Accessed 4 Aug. 2019

[17] “LHS-1 Lunar Highlands Simulant Fact Sheet.” UCF.

edu, Mar. 2019, http://sciences.ucf.edu/class/wp-content/

uploads/sites/58/2019/02/Spec_LHS-1.pdf

[18] McKay, David S. et al. “JSC-1: A New Lunar Soil Simulant.”

USRA.edu, 1994, https://www.lpi.usra.edu/lunar/

strategies/jsc_lunar_simulant.pdf

[19] Turkevich, A. L. “The average chemical composition

of the lunar surface.” NASA Astrophysics Data System,

http://adsabs.harvard.edu/full/1973LPSC....4.1159T. Accessed

17 May 2019.

[6] Miller J., et al. 2009. “Lunar soil as shielding against

space radiation”, page 167. Radiation Measurements

[7] Barry, Patrick L. “Radioactive Moon.” NASA.gov, 8

Sept. 2005, https://science.nasa.gov/science-news/science-at-nasa/2005/08sep_radioactivemoon

[8] “What is Radiation?” World Nuclear Association,

https://www.world-nuclear.org/information-library/safety-and-security/radiation-and-health/radiation-and-life.aspx.

Accessed 13 Aug. 2019.

[9] NASA Space Flight Human-System Standard: Volume

1, Revision A: Crew Health. NASA.gov, pp. 22, 30 July

2014. https://standards.nasa.gov/standard/nasa/nasa-std-

3001-vol-1. Accessed 22 Oct. 2019

[10] “Building the LRO Spacecraft.” NASA, https://lunar.

gsfc.nasa.gov/spacecraft.html. Accessed 3 Oct. 2019

[11] Litvak, M. L. et al. “Global maps of lunar neutron

fluxes from the LEND instrument.” Journal of Geophysical

Research, vol. 117, June 2012, pp. 1-2, https://www.

researchgate.net/publication/235979801_Global_maps_

of_lunar_neutron_fluxes_from_the_LEND_instrument.

Accessed 28 Oct. 2019

[12] Woolum, Dorothy S. et al. “The Apollo 17 Lunar

Neutron Probe Experiment.” Astrophysics Data System,

http://adsabs.harvard.edu/full/1973LPI.....4..793W. Accessed

28 Oct. 2019

80 | 2019-2020 | Broad Street Scientific PHYSICS


MODELING THE GRAVITATIONAL LENS OF THE

EINSTEIN RING MG1131+0456

Nathaniel S. Woodward

Abstract

We present the results of modeling the distribution of mass in the elliptical lensing galaxy of the Einstein ring MG1131+0456

using numerical and computational modeling methods. We applied two mass models of varying complexity to this system:

a point mass and a singular isothermal sphere. In constructing these models, we relied on image construction through

ray-tracing. Results from the ray tracing method qualitatively agree with both the literature and the observed radio imaging

of MG1131. Total mass derived from the ray tracing model was within the range predicted by previous models. Current

work aims to extend the ray-tracing model to include a singular isothermal ellipsoid as well as applying this model to

gravitational lenses that have not been analyzed.

1. Introduction

Gravitational lensing is a phenomenon predicted by

Einstein’s general theory of relativity where the path of

light is determined based on the curvature of the spacetime.

Gravitational lensing is a powerful tool when applied

to problems in modern physics and has allowed physicists

to measure universal constants such as the Hubble constant

and the deceleration parameter at a high degree of

accuracy, which quantify the expansion and deceleration

of the expansion of the universe, respectively [1][2]. Gravitational

lensing has also allowed physicists to determine

distributions of dark matter in galaxies and throughout the

universe and construct more accurate models of galaxies

[3]. In recent years, gravitational lensing analysis has been

conducted primarily through the comparison of observation

studies and computational studies that rely on mass

models to represent the lensing object. In this paper we

will examine the gravitational lens MG1131+0456 and

aim to calculate the mass distribution and total mass. We

will compare our results directly to those of Chen et al. to

verify our model [4]. After testing our model against the

literature, we hope to determining the total mass and mass

distribution of systems that have not been studied.

1.1 – MG1131+0456 as a gravitational lens

MG1131+0456 is an extensively studied gravitationally

lensed system and because of this we can characterize it

based on the previous literature. The literature suggests

that the source of MG1131+0456 is a distant quasar within

the range of 3500 to 4200 Mpc [6]. The radiation emitted

by this quasar is believed to be lensed by a massive elliptical

galaxy and the resulting images have been analyzed in

the infrared and radio wave wavelength [5][6]. The exact

orientation of this galaxy has been debated in the literature

and because of the uncertainty, we will work to examine

a variety of orientations in our model. Figure 1 displays

MG1131+0456 in the radio wavelength as recorded in the

CfA-Arizona Space Telescope LEns Survey of gravitational

lenses (CASTLeS) [7]. The image of MG1131+0456 in

the radio wavelength expresses two distinct features: a near

complete Einstein ring and two distinct compact images.

Previous studies have utilized these features to suggest the

source image may be composed of two compact components

[5]. One of these components is collinear with the

lens and the observer and is lensed into an Einstein ring.

The other compact component is slightly offset from the

line connecting the observer and lens and as a result, is

lensed into the two compact multiple images. We will conduct

our gravitational lensing analysis under the assumed

geometry of the source where the quasar is composed of

two distinct compact components.

Figure 1. The gravitational lens MG1131+0456 in radio

imaging. Credit: Hewitt et al. 1988. [5]

1.2 – Distance Values

The distance to the source has been debated throughout

PHYSICS

Broad Street Scientific | 2019-2020 | 81


the literature, but based on redshift data, we can construct

a range of possible distances. Work by Tonry et al. estimated

a source redshift ranging from z = 1.9 to z = 5 [6].

We assumed the predictions by Tonry et al. as bounds for

all possible redshifts for the source quasar. We first solve

the equation for recession velocity v from the relativistic

Doppler effect, shown in Eq. 1 where γ is the Lorentz factor

and z is the redshift.

From the range of redshifts, we use the source redshift

to calculate a recession velocity using Eq. (2).

(1)

(2)

After calculating a recession velocity, we can apply

Hubble’s law in Eq. 3 with a Hubble constant value of

69.8 km/s/Mpc from the results of Freedman et al. [8]

to determine the distance to the source.

(3)

Using Eq. 2 and Eq. 3, we computed a source distance

in the range of 3500 to 4200 Mpc. While the redshift of

the source is debated, the redshift of the lensing object

is agreed upon as 0.84 [6]. Using the same procedures,

we calculated a distance to the lens of 2400 Mpc [6]. We

will refer to the distance to the source and the distance

to the lens as D s and D l respectively.

Cartesian everywhere except the lensing plane, we assumed

that D s − D l = D ls . As shown in Figure 2, under

the thin lens approximation, this assumption holds.

2. Materials and Methods

2.1 – Preliminary Point Mass Model

Determining a mass model for a lensing object is an integral

part of the lensing analysis. First we conducted an

analysis modeling the lensing galaxy as a point mass. The

aim of this analysis was simply to validate that our distance

values from the literature for the lensing system were approximately

correct. In creating our point mass model for

the lensing galaxy, we set all distance values equal to the

values determined from the literature.

Examining the data from Hewitt et al. [5], shown in

Figure 1, we designated the two compact components as

multiple images and computed the Einstein ring radius.

The Einstein ring angular radius is a characteristic of a

lensing object and can be calculated for a point mass model

by Eq. 4, where θ E

is the Einstein ring angular radius and

θ+ and θ− are angular deviation of the positive and negative

multiple images respectively [9]:

The simplicity of the point mass model allowed us to

iterate through possible distances within the range of 3482

Mpc to 4183 Mpc and calculate a system mass using Equation

5:

(4)

(5)

Under the assumption that D s − D l = D ls , we can

reduce Eq. (5) to;

(6)

Figure 2. We can see that under the thin lens approximation

where the lensing object is represented

as a plane, our distance approximation holds. In

addition to displaying distance values, this diagram

also shows the angular values that describe a lensing

system, where θ is the angle of observance, β is the

angular deviance of the source from the optical axis,

andˆα and α are the deflection vector and reduced

deflection angle, respectively. Credit: [9]

We also define a variable D ls that represents the distance

between the lens and the source. Under the thin

lens approximation, which assumes spacetime to be

By nature of the point mass model, there exists a possible

mass such that at every point within the source

distance range an Einstein ring and multiple images

would form to match the data. This yields a range of

possible masses calculated from the bounds of the distance

range. With a known Einstein ring angular radius

and lens distance, we are left one free parameter of the

source distance. We calculate the total mass of our point

mass model using Eq. 6 by iterating through source distances

derived from Tonray et al.’s redshift predictions.

Our analysis predicted a mass in the range of 2.9 * 10 41

kg to 4.0 * 10 41 kg. A ray tracing diagram of our point

mass model is shown in Figure 3 which displays the distance

range and shows light rays emitted from two distinct

compact components. Our predicted values from

the point mass model are within ≈ 20% compared to the

result of 5.1 * 10 41 kg predicted by Chen et. al. [4].

82 | 2019-2020 | Broad Street Scientific PHYSICS


Figure 3. The ray tracing diagram above displays

how the literature suggests the lensing system

MG1131+0456 is composed. There are two luminous

points of the source quasar representing the compact

components of the source. Rays of light travel from

these points to the lensing plane and some of these

rays are deflected to reach the observer. The diagram

also displays the minimum and maximum distance

values for the source quasar which are labeled ”Min”

and ”Max.” Dashed lines correspond to locations of

virtual images for the multiple images.

2.2 – SIS Mass Model

After completing our rudimentary analysis using the

point mass model, we extended our model to describe an

SIS mass distribution. A Singular Isothermal Sphere model,

or SIS, is a simple, but effective model for gravitational

lensing systems. An SIS model represents a galaxy as a

sphere of gas in hydrostatic equilibrium[9]. In this model,

the motions of particles in the spherical gas cloud in

hydrostatic equilibrium are analogous to the orbital motion

of objects in the galaxy. Equations that describe an

SIS model are derived directly from the ideal gas law and

the equation of hydrostatic equilibrium [9]. The SIS model

and its deflection of light rays are shown in Figure 4.

The SIS mass model allows us to include a non-uniform

mass distribution which is fundamental to an accurate approximation

of a galaxy. Every parameter of the SIS mass

model depends on the velocity dispersion or σ. The velocity

dispersion is analogous to the temperature of a cloud of

particles in hydrostatic equilibrium, but for the motion of

planets and stars in a galaxy, it describes the characteristic

speed of masses in motion [9]. The velocity dispersion can

be treated as a free parameter, but often is estimated to

225 km/s. With the velocity dispersion known or parameterized,

we can calculate properties of an SIS mass model

using the equations [9]:

(7)

(8)

The mass within radius R or M(R) is a function of the

impact radius R, which is the radial distance to the center

of lens from the point a ray of light impacts the lensing

plane, and grows without bound as R approaches infinity

(shown in Eq. 7). The mass density within a radius R or

Σ(R) is inversely proportional to R and therefore predicts a

singularity of infinite density as R approaches zero (shown

in Eq. 8). By limiting the upper bound of R to some radial

extent, we ensure that the total mass is finite, but a finite

core density can only be achieve through extending to a

more complex mass model. A key characteristic of the SIS

model is shown in Eq. 10, a deflection angle independent

of R.

(10)

Figure 4. A source object of varying brightness is

shown to be emitting light which impacts the lensing

plane at arbitrary points. We can see that the

lens plane contains a spherical lensing object which

deflects the incident rays of light. Some of the deflected

rays of light impact the observer and become

images of the source.

2.3 – Ray Tracing SIS Model

We chose to devise a new method to characterize our

lensing system by tracking the path of individual light

rays as they impacted the lensing plane. We predict an

image based off the light rays calculated to impact the

observer. In creating this model, we chose to work primarily

with the SIS model as it reduces the complexity

of the ray tracing model and is shown to be a reasonable

model for an elliptical galaxy. The SIS lowers complexity

because it is spherically symmetric, so to compute

light deflection we only require the impact radius.

First we defined a source matrix where every cell

has an assigned unitless brightness value. We then convert

these brightness values into pixels by converting

all brightness values into percent maximum brightnes.

The source matrix that we use to construct the source

plane and the corresponding source plane are shown in

Figure 5.

(9)

PHYSICS

Broad Street Scientific | 2019-2020 | 83


lens (x l , y l , z l ). We defined z s = −D ls and z l = 0. We use

these two points and the center of lens (0, 0, 0) to define

the plane that will contain the entire path of our deflected

light ray. We will first rotate about the z-axis such that the

impact point (x l , y l , z l ) has an x-component of zero (rotating

the y-axis to the impact point).

(11)

Figure 5. We initially define the source numerically

with a matrix having cell values corresponding to

point sources of light. We determine the brightness

of each point source of light based on the ratio to the

maximum brightness in the source matrix.

Similar to our construction of the source plane, we define

a lensing plane with coordinates (x l

, y l

) to represent

our SIS and use the origin of our lensing plane as one of

the two points to define the optical axis (with the other

point being the observer). To model light rays traveling in

all directions from the source, for every luminous point on

the source plane, we extended a light ray to every possible

point on the lensing plane. We can determine the incident

light ray directly from its lens plane and source plan coordinates

using simple vector algebra.

To allow light rays to travel between the source plane

and the lensing plane, we extended our coordinate system

to a three dimensional space. We assign each plane in

three-dimensional space a constant z-coordinate. This assumes

that our lens and image planes are parallel. We hope

to incorporate tilt in the source plane in further work.

In designating the z-coordinates of our planes, we again

based our distance values in the literature. The distance to

the lens was found to be 2400 Mpc and, as stated in our discussion

of the point mass model, the distance to the source

is a variable range. We chose the lower bound of the range,

3500 Mpc, initially to work with for our distance to the

source. With these values from the literature and the use

of our conversion function, we successfully combined our

source and image planes into one three dimensional space.

2.3.1 – Coordinate Manipulation Vector Verification

After creating the shell of our lensing model, we implemented

a method to verify if a light ray impacted the

observer by calculating a deflected vector to represent the

light ray and determining if this deflected ray impacted the

observer. While this method is computationally intensive,

it maintains the needed level of complexity to describe the

system.

First, we aimed to define a plane that contains the entire

path of the light ray, thus reducing the dimensions by one.

We achieved this through a series of coordinate rotations.

As previously stated, we can define the path of light between

the source point (x s , y s , z s ) and impact point on the

Next, we perform another coordinate rotation to

rotate the source position on to the y-axis, therefore allowing

the entire path of our light ray to be described

the only the y and z components. We will rotate relative

to the y-axis at an angle θ z . This rotation is shown in

Figure 6b.

(a) z-axis rotation

(12)

(b) y-axis rotation

Figure 6. (a) We extend a ray of light from an arbitrary

point on the source plane to an arbitrary point

on the lens plane. We then rotate the coordinate

system about the z-axis by an angle θ y to align the

impact location with the y-axis. (b) Viewing the system

from top-down, we rotate the coordinate system

about the y-axis such that the source location also

has an x-component of zero. Having both the source

and the impact vectors with a zero x-component vector

allows us to define the incident path of light in

two dimensions.

After these two coordinate transformations, we can reduce

our three-dimensional vectors for the source location

and lens location on the lensing plane to two dimensional

vectors with components (z, y). Our lens and source vectors

can be represented respectively as:

84 | 2019-2020 | Broad Street Scientific PHYSICS


We can determine the components of a ray starting at the

source and pointing to the impact point on the lens by subtracting

the source vector from the impact vector. Therefore

our incident vector I is,

Because under the SIS model, the deflection angle α is

constant, we can define a deflection matrix G which will

represent our lensing galaxy. Matrix G acts as a rotation

matrix and deflects the incident vector byˆα. This matrix

is shown below and the vector deflection is represented in

Figure 7.

deflected vector is parallel to the ID vector then it will impact

the observer and an image will be formed. To check if

the deflected vector is parallel to the ID vector, we verify

that the ratio between the y and z components are equal

for both the deflected and ID vector. We recognized the

limitations of a matrix representation for both the source

image and lens plane and incorporated an allowed variance

in the y-component. The best value for allowed deviance is

still being determined.

After determining that an image would be formed, we

increase the brightness value at (x l

, y l

) by the source image

brightness at the point S(x , y s

). We repeat this process for

s

all luminous points on the source plane, thus checking for

image formation for every possible ray of light from the

source.

So far, we have worked under the assumption that the

center of our source plane lies on the optical axis. The deviance

of a source image from the optical axis is referred

to as β and is shown as an angular value in Figure 2. The

literature suggests that the source of MG1131+0456 is

composed of two distinct compact components: one on the

optical axis which is lensed into an Einstein ring and one

that deviates from the optical axis and is multiply imaged

[5]. Assuming that the images shown in Figure 1 are in

fact multiply imaged, we can calculate an angular deviance

using SIS lensing equations [9].

(13)

(14)

Figure 7. Having reduced the path of light to a two

dimensional plane, we apply the deflection matrix

G(α) to deflect the light ray.

To reduce needless computation, before calculating the

deflection vector, we check if the observer exists within an

allowed variance from the plane that describes the path of

the ray of the light. We know that the path of the incident

ray and the deflected ray are entirely contained in the plane

defined by the source, the center of the lensing object, and

the impact point on the lensing object. The observer is not

in this plane or within a small range, then it is not possible

for the light ray to be seen. The best value for allowed deviance

is still being determined. If the observer is in the plane

of the ray’s path, then in order to calculate the deflection

vector, we we must multiply the deflection matrix and the

incident vector:

To verify if the deflected vector impacts the observer, we

define an ID vector that starts at the impact location on

the SIS and points to the observer. We know that if the

Through summing these equations, we can solve for Beta

to be:

(15)

Using the data from Hewitt et al. we applied Eq. (15) to

calculate a β value of 2.8x10 −6 radians [5]. From an angular

deviance, we use simple trigonometry to calculate linear

deviance in the source plane using the relation, where D β is

deviance vector in the source plane:

(16)

Having calculated a deviance in the source plane, we recognize

that the SIS model is axially symmetric and any

direction of deviance will be equivalent in its effect. We

designate an arbitrary direction of deviance relative to the

x-axis of π/4 radians and with this arbitrary direction, we

can determine the x and y components of deviance. These

components are shown in Eq. 17 using the small angle approximation

for tan(θ) where θ D

is the angle of deviance

and D x

and D y

are the x and y components of deviance respectively:

and (17)

PHYSICS

Broad Street Scientific | 2019-2020 | 85


With x and y components of our deviance, we can

simply add our respective components to all x and y coordinates

of images in the source plane and thus shift all

luminous points by the deviance. This process is shown in

Figure 8.

we designate the velocity dispersion of our SIS model as a

free parameter. We require the velocity dispersion to be a

free parameter because it is one of the largest contributors

of both the total mass and mass density, both of which are

values we aim to calculate. We have shown through Eq.

(7) that the SIS total mass equation requires a radial limit

and because of this we allow the radial limit of the lensing

SIS to vary within the range of 10 5 pc to 2x10 6 pc. We have

calculated the angular deviance between the two source

components to be θ D

= 2.8x10 −6 radians (Fig. 8).

2.6 — Calculating Variance

Having defined free parameters, before predicting results,

we must construct a method to quantify the goodness

of fit for our predicted image. We recognize that the

source data are pixelated just as our predicted data, but

their respective resolutions are not necessarily the same

(Fig. 9). We aim to reduce each resolution to a 100x100

image such that we can directly compare individual pixel

values to calculate a source variance.

Figure 8. In the figure, we can see that the source is

originally aligned with the optical axis. We allow

for a source offset from the optical axis by shifting

the source plane with a deviance vector D which we

can decompose into its components D x and D y . With

these components, we can transfer every coordinate

in the aligned source plane to the deviated source

plane by adding D x and D y to the x and y coordinates,

respectively.

2.4 — Modeling the Source Quasar

To begin testing our lensing model, we must construct

an accurate representation for our source. The literature

suggests that the source is likely a distant quasar with two

extended compact components embedded in a galaxy [10].

Our initial method of modeling assumes a symmetric distribution

of matter in the quasar. However this representation

resulted in a needless increase in computational load

compared to the results of the point source model. We determined

that a point source model for the source quasar

served as the most effective model.

2.5 — Free Parameters of the Ray Tracing Model

In order to apply our model to MG1131+0456, we must

designate free parameters. From previous work by Tonry

et al. [6], there is a firm lower bound at redshift z s

= 1.9

and an estimated upper bound at redshift z s

= 5.0. Previous

findings suggest the source distance as a viable free

parameter as it is uncertain and constrained. Additionally,

Figure 9. We can reduce a pixelated portion of an image

to one individual pixel with a brightness equal

to the average of all of the preexisting pixels. In this

diagram, we reduce an image with a pixel ratio of 4:1,

as apparent because we average the brightness values

of a 4x4 square of pixels.

First, we determine the ratio of pixels between the current

image and the ideal 100x100 image. For example, if

our predicted image was 400x400, we would have a pixel

ratio of 4:1 from the predicted to the ideal. Using this, we

average every 4x4 square of pixels, which do not overlap,

in the predicted image to form one pixel in the reduced image.

This process is shown in Figure 9. Additionally, background

noise is apparent throughout the image which adds

inherent variation when compared to our computational

model. To resolve background noise, we apply a lower

bound to pixel brightness which filters out noise while retaining

the structure of MG1131, shown in Figure 10a.

3. Results

The best fit model for MG1131 using the ray tracing

model resulted in a velocity dispersion of 255 km/s,

86 | 2019-2020 | Broad Street Scientific PHYSICS


(a) Reduced Astronomical Data (b) Best Fit Model (c) Overlap Image

Figure 10: (a) Using the methods previously described, we have reduced the radio image of MG1131 to a

100x100 pixelated image where pixel brightness again correlates to percent maximum brightness. This

image allows for direct pixel to pixel comparisons between the ray tracing model and the data. (b) Using the

ray tracing model we determined the best fit model compared to the reduced and noise-filtered radio imaging

of MG1131 using parameters of source distance, velocity dispersion, radial extent of the lensing SIS, and

the angular deviance of the source components. (c) By overlapping the best fit image and the observed data

of MG1131 we better understand the limitations of the ray tracing model. The observed data for MG1131 are

shown in red, the computed image is shown in green, and all overlap is yellow. We are unable to represent

the ellipticity that the observed image contains which significantly limits the accuracy of our model. Additionally,

the best fit model does include multiple images, but they are not aligned with the multiple images

present in MG1131.

a source distance of 3482 Mpc, a β deviance of 2.8x10 −6

arcseconds, and a lensing radius of 12500 pc. This simulation

was run using a SIS with a radial extent of 25000

pc. Using these parameters and Eq. (7) we calculate a total

mass of the lensing galaxy to be 6.98x10 11 M sun

. Comparing

our results to the work of Chen et al. [4] which predicted

values in the range of 1.17x10 12 M sun

to 2.57x10 11 M sun

, our

best fit model is within this range. Our predicted mass using

the ray tracing model varied by 46% and 92% compared

to the least and most massive models presented by Chen et

al. However, compared to the average of the values predicted

by Chen et al., 7.14x10 11 M sun

, there is a percent differnce

of 2.2%. The best fit model is shown in Figure 10.

4. Discussion and Future Work

The best fit model using the ray tracing method qualitatively

fits the radio image of MG1131 having an Einstein

ring and two distinct multiple images of the offset source.

However, due to the limitations of the SIS model when

applied to an elliptical lens and the assumption of point

sources, we cannot represent all features of MG1131 in

this model. While the ray tracing model cannot reflect the

complex optical strucutre of MG131, our results are within

the range of values predicted by Chen et al. [4] which suggest

that our model may be able to predict a realistic mass

for complex gravitational lenses.

In future work, we aim to further apply the ray tracing

model to previously analyzed systems to test its mass predictions

against those of the literature. We are first examining

the system B2045 +265 shown in Figure 11.

Previous work by Fassnacht et al. concluded that B2045

is "a radio galaxy lensing a radio-loud quasar" [11]. We

have shown through the comparing mass predictions for

MG1131 with those of Chen et al. [4] that the ray tracing

method is viable for modeling strong lensing events

by galaxies. The similar structure of B2045 leaves it as an

optimal candidate to retest the ray tracing model’s accuracy

in mass predictions. Following the literature, we plan to

model B2045 with an SIS model and a source represented

as multiple point sources [11].

5. Conclusion

The ray tracing model has shown to be a viable method

for computationally modeling gravitational lenses but has

several limitations at high levels of complexity. We have

shown that regardless of these complexity limitations, the

ray tracing model predicts a realistic mass model for the

gravitational lens MG1131 through comparing our results

to the literature [4]. Future work aims to extend the ray

tracing model to include a singular isothermal ellipsoid as

the mass model to allow for ellipticity in lensed images as

well as applying the ray tracing model to B2045 as well as

unanalyzed systems.

PHYSICS

Broad Street Scientific | 2019-2020 | 87


org/10.1086/175857.

[5] Hewitt, J. N., et al. “Unusual radio source MG1131

0456: a possible Einstein ring”. In: Nature 333.6173 (1988),

pp. 537–540. doi: 10.1038/333537a0.

[6] Tonry, John L. and Kochanek, Christopher S. “Redshifts

of the Gravitational Lenses MG 1131+0456

and B1938+666”. In: The Astronomical Journal 119.3

(Mar. 2000), pp. 1078–1082. issn: 0004-6256.doi:

10.1086/301273. url: http://dx.doi.org/10.1086/301273.

[7] C.S. Kochanek, E.E. Falco, C. Impey, J. Lehar, B. Mc-

Leod, H.-W. Rix. “CASTLES”.

Figure 11: Images of B2045 from the CASTLeS lensing

database and aim to construct a rudimentary model

of the system using the ray tracing model. Credit: [7]

[11]

6. Acknowledgements

I would like to thank Dr. Jonathan Bennett from The

North Carolina School of Science and Mathematics for

guidance throughout the research process as well as suggesting

the field of gravitational lensing. Additionally, Dr.

Bennett’s advice was crucial in designing the ray tracing

method. I would also like to thank The North Carolina

School of Science and Mathematics and The North Carolina

School of Science and Mathematics Foundation for providing

me with this opportunity and the funding necessary

to produce this work.

[8] Freedman, Wendy L., et al. “The Carnegie-Chicago

Hubble Program. VIII. An Independent Determination of

the Hubble Constant Based on the Tip of the Red Giant

Branch”. In: The Astrophysical Journal 882.1 (Aug. 2019),

p. 34. issn: 1538-4357. doi: 10.3847/1538- 4357/ab2f73.

url: http://dx.doi.org/10.3847/1538-4357/ab2f73.

[9] Congdon, Arthur B. and Keeton, Charles. Principles of

Gravitational Lensing. Springer International Publishing,

2018. isbn: 978-3-030-02122-1. doi: 10.1007/978-3-030-

02122-1.

[10] Kochanek, C. S., et al. “The Infrared Einstein Ring in

the Gravitational Lens MG J1131 0456 and the Death of

the Dusty Lens Hypothesis”. In: The Astrophysical Journal

535.2 (2000), pp. 692–705. doi: 10.1086/308852.

[11] Fassnacht, C. D., et al. “B2045 265: A New Four-Image

Gravitational Lens from CLASS”. In: The Astronomical

Journal 117.2 (1999), pp. 658–670. doi: 10.1086/300724.

7. References

[1] Deceleration Parameter: COSMOS. url: http://astronomy.swin.edu.au/cosmos/D/Deceleration%20Parameter.

[2] Huchra, John P. url: https://www.cfa.harvard.edu/~dfabricant/huchra/hubble/.

[3] Blandford, R. “Cosmological Applications of Gravitational

Lensing”. In: Annual Review of Astronomy and

Astrophysics 30.1 (Jan. 1992), pp. 311–358. doi: 10.1146/

annurev.astro.30.1.311.

[4] Chen, G. H.; Kochanek, C. S.; and Hewitt, J. N. “The

Mass Distribution of the Lens Galaxy in MG 1131+0456”.

In: The Astrophysical Journal 447 (July 1995), p. 62.

issn: 1538-4357. doi: 10.1086/175857. url: http://dx.doi.

88 | 2019-2020 | Broad Street Scientific PHYSICS


AN INTERVIEW WITH MR. ERIK TROAN

From left, Jason Li, BSS Editor-In-Chief; Eleanor Xiao, BSS Publication Editor-In-Chief; Dr. Jonathan Bennett, BSS

Faculty Advisor; Olivia Fugikawa, BSS Editor-In-Chief; Megan Mou, 2020 BSS Essay Contest Winner; and Mr. Erik

Troan '91, founder and CTO of Pendo.io.

To get us started, you are a self-proclaimed technology

geek. When was it and what was it that first got you

interested in computer programming and engineering?

I’ve been doing [CS and Eng] since I was probably five or

six years old. Going back a few decades ago, people didn’t

have computers at home, but my father was at IBM so he

was always around them. He put me in a programming

class at the library when I was probably six or seven. We

also had a little computer at school so those two things

made me kind of interested in exploring the world. I know

what I enjoyed about it then is probably what I still enjoy

about it now; it’s a combination of problem solving while

doing it in an abstract environment where you really can

control a lot of things. Many of you probably enjoy math.

Math is fun because there is a right and a wrong, correct?

You get to solve a problem. You’re solving a puzzle, but

you’re not constrained by inconveniences like physics,

gravity, forces, or having to go build things and buy equipment.

For me, growing up, it was always a way I could

build and explore things. I could do things that didn’t have

the constraints of the physical world or have to get materials

and put them together. Physical things break and then

you have to fix them. So I think that level of control over

computer software is really what made me interested in it.

You’ve highlighted NCSSM as one of the most valuable

parts of your education. What was your favorite part of

your experience there?

The other people. It’s all about the kids. It’s about the social

scene—and I don’t necessarily mean social like parties, but

social in terms of, you’re in a peer group that’s smart. I’ve

probably said something like that at Convocation: you’re

not going to sit in a group of people as smart and as energetic

and as intellectual and as curious as you are ever

again and that leaves a mark. You really do realize there

are like-minded people out there; you have interesting

conversations and you’re solving interesting problems. I

remember I took a fractal class senior year and just the excitement

of the kids getting together to solve the problem,

and to really learn about it, is what really made an impact

on me.

I learned the importance of communication, so I learned

to write. I had some very good instructors. I learned from

FEATURED ARTICLE

Broad Street Scientific | 2019-2020 | 89


them that writing was something that helped you communicate,

and if you want to become effective in this world,

you’re not going to do it mostly one-on-one; you’re going

to do it one-to-many. Even in the world of the Internet,

Youtube and everything else, writing is probably the most

basic way that you’re going to communicate to a lot of people.

I also really fell in love with history, of all things. I had a

great U.S. History teacher. That’s followed me my whole

life. I’m still a history buff; I read history books all the time.

I was so fortunate, just like you, to have instructors who

pushed me and made me better all the time. I have a son

who is [in NCSSM Online], and it’s great watching him

online and how he kicks me out of his room so he can do

his classes because he’s getting that extra push and really

interesting problem-solving material. He just started his

new Bioinformatics class this week and you can just see the

lights back on in his eyes.

What do you do at Pendo now? What does a day look

like for you?

There are two different parts to that question. My official

job at Pendo is that I’m Chief Technology Officer, and that

I run all of Product Engineering. All the products that we

build come out of my team; it’s a team of about 110 people

now. [Our product] runs in the cloud (out on the internet)

because it’s a software product. The team deploys that

product, keeps it running, manages automated testing, and

does 24/7 monitoring. All of that is all under me. So [my

job is to] get the product, work with the team that defines

what we should build, figure out how to architect it, build

it, get it shipped to customers, and keep it up and running.

What do I do every day? I go to meetings pretty much all

day—the most important part of my job is helping with

communication within my team or across teams. I don’t

make the big decisions. If I’m making a decision, there’s

probably something wrong. It’s much more about getting

the right people in the room, enabling them to make the

decision and make sure they feel supported and empowered

to make the decision. Asking questions is also a big

part of it. I actually sat in your chair a few hours ago going

through whether or not we should use a third-party partner

for something new we’re rolling out. There was a team

of three people that told me two or three weeks ago that

they were really excited about it. Now they all changed

their mind and all I did was ask them questions about their

decision. “Why?” “Have you thought about this?” “What

about that?” I didn’t make the decision though. We were

going with their recommendation, but I helped to validate

and make sure they had the confidence that they thought

through everything. So I do that with both Engineering

and Product. The other thing I’ll do is work with Sales,

Product and Marketing to try to understand the market.

I’ll talk to customers about what Pendo does for them and

where they are having a problem. I think I have three or

four customers that I’m an executive sponsor for, so another

part of my job is making sure that they are successful.

It’s a lot of talking. I joke that the CTO is actually Chief

Talking Officer; some technical in it, but not much.

One thing in a fast-growing organization, you’ll find, is the​

problems d​on’t change; the problems change scale​. When

you solve a problem for a team of 20 with the right solution

for 20, when you get to a [team of] 100, that solution

probably doesn’t work any more. For example, you might

realize that you didn’t have a process for something at 20.

That didn’t matter at the time because two people went to

lunch once a week, saw each other informally at lunch and

communicated, “That’s not working,” or “Fine, I’ll fix it for

you.” Then they go back and they fix it. When you get to

100 people, those two people could now sit on different

floors and don’t go to lunch any more. All of a sudden,

you have a hole where things that were solved informally

can’t be solved informally any more because the roles that

you need to have around the table drift apart. That doesn’t

mean you screwed up, or that your process is wrong, it just

means that you’ve grown a lot, and you have to make sure

you’re constantly looking for things where you missed or

where you can do better. So as you hire specialists, they can

thrive without there being gaps between all the specialists.

Do you miss the coding part of your job or the engineering

part of your job?

No, because I still do it. This is one of those things where

it’s either the best thing about me or the worst thing about

me as the CTO. I’m still technical and still want to code.

It means I have to hire and surround myself with people

who complement that aspect of me. I have two VP’s of Engineering

who don’t code. So, I brought people in to take

things that gave me a little more room to still be technical.

[But] I have to be very careful with what I code. I tend to

code things that are longer term, things that aren’t going

to be blocking a customer, or things that are not a promise

that we are going to get it done by a certain day. I’ve been

told that most of the team loves that I [code], because it

makes me relatable and I can understand what they’re going

through a little more. But it also does mean that there

are other parts of my job that I delegate and I do differently

because I insist on continuing to code.

What is your advice to students who want to learn

more about coding?

Go to school. Go to college. I’m a huge believer in formal

computer science education. I hire people who don’t have

it, so I do hire coding academies and people who have just

90 | 2019-2020 | Broad Street Scientific FEATURED ARTICLE


picked it up, but the easy path to success if you want to be

in a computer software job is to go to college in computer

and software. You will learn things in school that you

won’t learn anywhere else. I took a lot of graduate classes

as an undergraduate and I still use those lessons. Some of

the best engineers have Master’s degrees. I haven’t hired a

Master’s degree who’s not a good engineer. These things

go together. People get smart in graduate school and undergraduate

for a reason, so don’t let that go.

While you’re [in school], do as many internships, real

world projects, volunteer projects, and capstone projects

as you can. You’ve got to go solve problems. You’ve got to

exercise the muscles in your brain to let you try different

ways of approaching problems, because there are all sorts

of different ways to do it. Unless you’ve done it 20 times,

you’re not going to know which [solution] is the best. College

and graduate school absolutely accelerate that process.

What is your advice to students who hope to start their

own business?

To start a company, you’ve got to find a problem that you

are passionate about, not a solution you are passionate

about. I tried a [starting a company with a] solution that I

really liked and that company fizzled and failed because it

wasn’t a ​problem.​Find something you think is worth solving.

Find c​ustomers​who think it’s worth solving—ideally,

before you build it. You can’t always do that, but if you

can find some way to get people who will validate this as a

problem worth paying for, there’s nothing to replace that.

One of the reasons I got really attracted to Pendo was, our

CEO gave me a problem that I understood, was excited

about and resonated with my career. He started talking to

potential customers who said, “Oh yeah! Talk to this guy,

they said this.” So he could clearly come in with a head start

towards, “Was this a valuable problem to solve? What was

important about the problem? How is this really going to

help customers?”

Another thing I would say is: it is very easy, and I’ve been

guilty of this in my career, to take customers for granted.

They are going to make you successful. The only things

that make you successful are their willingness to pay and

your willingness to make t​hem​successful. So don’t take

them for granted, especially if you’re talking about low

prices and products. It’s very easy to say, “Yeah, we have to

make it up in volume.” or “If any one customer is unhappy,

we can kind of blow it off.” If one customer is unhappy, it’s

probably a thousand customers who are unhappy. Make

your customers happy. Your customers will carry you everywhere.

We have been very lucky here. We’ve done five

venture-funding rounds. Every one of them was preempted,

so we never went to look for money. Investors came

to us, wanting to put money in. The reason that happened

FEATURED ARTICLE

was our customers. The first one that found us was because

we had already sold to six other portfolio companies,

and they started hearing about Pendo, Pendo, Pendo.

“What is this thing? Why are you guys all using it?” They

were so happy with the value we were delivering that they

told their investors about it, who told their partners about

it. It happened again and again. It wasn’t any magic. We

have seven core values, and one of them is “maniacal focus

on the customer.” We are absolutely dedicated to them. I

think out of everything I’ve learned, that is the one thing

that I have discounted in my career, and I wish I hadn’t. It’s

so important.

Many of the students here at NCSSM view failure as

something shameful. How did you overcome the failures

that you encountered in your life, and how did you

learn from them?

Practice. I used to be a private pilot. There’s a saying

around that, “The only way to avoid mistakes is through

experience, and the only way to get experience is by making

mistakes.” Failure is a side effect of taking a chance,

taking a risk, pushing your zone. If you don’t fail, you’re

not trying very hard. And I get it—I watched my kids go

through middle school and high school. I get that we built

a society for students where failure can be very, very expensive.

You guys get that one “C” in a class, and there’s

colleges you might not get into or scholarships you might

not get. You’re in a weird place in high school. Don’t mistake

how you succeed in that microcosm for how you’re

going to be successful in life. You’re going to be successful

in life by growing, learning, and trying new things. One

of the reasons that [the US] is still one of the premiere

economies in the world is that we don’t punish failure. I

failed in my last company. It was a big crater in the ground,

burned through tens of millions of dollars of other people’s

money. When I helped raise money for Pendo, people saw

that as a ​plus.​They didn’t say, “Well, your last one failed,

why would I invest in you now?” They said, “Hey, what’d

you learn?” and I said “This and this and this, oh my god

I learned so much,” and they’d say, “Great, let’s do this.”

You’re lucky to be in a country that, as you move out of

that narrow, high-achieving academic track that you’re on,

failure won’t always be punished. So embrace it and try

things. You have to try new things or you won’t get better.

I have a really smart PhD [employee] here, who worked

on a project for 6 months. He spent 6 months of engineering

on something. It didn’t work. At the end of the day,

we had to shelve it. But if we didn’t try, we would always

be wondering “What if?” We had to figure out other ways

to solve it, and that approach was a bad one. So, I get why

kids at Science and Math and in high school are scared of

failure; I think you’ve been taught your whole lives to be

scared of failure. Just try and get over it when you’re in college.

It’s going to be very hard to do anything special unless

Broad Street Scientific | 2019-2020 | 91


you take some gambles. You can’t win every gamble. But

when you fail, look at why. I learned more from the company

that didn’t work than from the companies that did.

I’ve hired people who have been fired, by the way. The

question is, “What did you learn from your failure? Why

did it not work? What was it about you that had to change?”

and if they say “it was the company, there were bad people

there, they didn’t listen to my ideas, they did everything I

told them not to” — run away. That’s not learning. If they

say “Yeah, the company had this culture, and I brought this

culture, and I had trouble adapting,” or, “I really thought

this was gonna work out, and I pushed hard for it. I think I

alienated some people” — those are the kind of people you

want, because they’ll probably be better next time.

What do you enjoy doing outside of work?

I scuba dive. I’m actually a scuba instructor, so I teach one

or two classes a year. I’d like to get back into flying; I’m

looking into doing that in the next year or two. I’m a pretty

avid snowskier; I try and go out west a couple times a year.

I run hundreds of miles a year—that’s not something I enjoy,

I do it because I’m old. I like to cook. It’s the highlight

whenever we can get the whole family around the table. I

have 3 teenage boys, so it doesn’t happen that often, but

when we can all sit down to a meal, our meals tend to last

one and a half, two hours. They’re very long, drawn-out

meals after the food’s all gone. We don’t see each other that

much so we enjoy that. I like to travel. Whenever I can,

I’ll take a trip overseas—not for work, but if I can travel

personally, it’s great to just be in a new place and wake up

somewhere else. I also like to read. I read a t​on.

As a scientist who works in the entrepreneurial side of

things, what do you have to say about the role of entrepreneurship

in advancing science?

First of all, I’m not a scientist, I’m an engineer. I consider

those really, really different. When I didn’t go to grad

school, I turned away from being a scientist. Scientists discover

new things. Engineers try and apply science to problems.

I think engineering and entrepreneurship are really

tightly coupled. My personal belief is that most of the time

if you engineer something new, you probably need a new

company to really make it successful. The iPhone is a huge

counterexample to that, but most truly novel engineering

projects have a company to go off of because [they’re]

bringing something new to the market.

research grants and universities where you don’t need to

have that immediate payback period.

Some of the most interesting things that come out of science

start as an intellectual curiosity. They didn’t start off

as solving a real-world problem, but they all feed into real-world

problems, into the hands of people like me, who

are engineers. When you start crossing those wires too

much, you get into bad positions like you have in medicine,

where a lot of research in medicine is being privatized.

If it isn’t for a disease that has enough payback, then

it gets shelved, so you start getting the wrong incentives

around basic research.

How important is entrepreneurship for applying the

knowledge that comes out of science, or for doing engineering?

I think for building real solutions to real problems, it’s critical.

If all we had in the world was fifty Fortune 50 companies,

then we wouldn’t get innovation, and we wouldn’t

get new products to solve new problems out there. The

easiest example is alternate energy. It’s not Exxon and BP

that are bringing alternate energy and clean energy to

the world. They have a business—they dig up oil, refine

it, burn it, and make a lot of money doing it. It’s hard to

take a business like that and go do something that’s new

and different. It’s a risk. It’s a lot easier to take a risk as an

entrepreneur because you have a lot to gain, but if you’re

running a big company, how do you value that risk? You

have a lot to lose.

[Entrepreneurship] is critical. I think universities are

normally pretty bad at it. Stanford is uniquely good at it,

and I don’t know if it’s [because they’re] Stanford or if

it’s because they’re sitting in the middle of Silicon Valley

and have hundreds of billions of investment capital sniffing

around looking for the next opportunity all the time.

Mostly it’s disappointing when you look at universities

that have so much pressure to raise money and own part

of their inventions that they aren’t good at getting these

companies up or bringing the technology up. Which is too

bad because I think there’s a huge amount of symbiosis between

those two.

Science is harder. A lot of science has around 30 or 40

year payback periods, and it’s not that they’re not interesting

topics and that they’re not important. If you’re lucky

enough to be a scientist, and you’re able to do it in an entrepreneurial

sense, I think that’s great. But for most science,

I think, you need to figure out how to do it through

92 | 2019-2020 | Broad Street Scientific FEATURED ARTICLE



Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!