Professor’s book about Sephardic Jews chosen as a best of 2019

Photograph of Sarah Abrevaya Stein

Sarah Abrevaya Stein. Photo credit: Caroline Libresco

Adding to the chorus of critics’ raves, The Economist has named “Family Papers: A Sephardic Journey Through the Twentieth Century,” a new book from UCLA professor of history Sarah Abrevaya Stein, to its best of 2019 list.

Stein’s latest work explores the intertwined histories of a single family (the Levys), Sephardic Jewry, and the dramatic ruptures that transformed southeastern Europe and the Judeo-Spanish diaspora. It has received glowing reviews from The New York Times, the Wall Street Journal (subscription required), the Los Angeles Review of Books and more.

Stein, who holds the Maurice Amado Endowed Chair in Sephardic Studies at UCLA, spent a decade researching this work, a journey that took her to a dozen countries, dozens of archives, and into the homes of a Sephardic clan that constituted its own, remarkable global diaspora.

The phrase “Sephardic Jew” refers to those of Spanish or Hispanic background. Stein’s new book begins with a family originally from old Salonica, a Mediterranean seaport of the Ottoman Empire, now Thessaloniki, Greece. In the late 19th century it was home to a large community of Spanish Jews.

The idea to tell the story of the Levy family came as Stein researched another book, an English-language translation of the first Ladino (which refers to a background of mixed Spanish, Latin American or Central American heritage) memoir ever written, “A Jewish Voice from Ottoman Salonica: The Ladino Memoir of Sa’adi Besalel A-Levi.”

“He spent the last years of his life writing a Ladino-language memoir to air a lifetime’s worth of grievances,” said Stein, who is also the Sady and Ludwig Kahn Director of the Alan D. Leve Center for Jewish Studies at UCLA.

Her book traces the history of a collection, how one family archive came to be built and preserved.

The existence of Sa’adi’s personal memoir, one sole copy written in written in soletreo (the unique handwritten cursive of Ladino), outlived wars; the collapse of the empire in which it was conceived; a major fire in Salonica; and the Holocaust, during which Jewish texts and libraries as well as Jewish bodies were targeted by the Nazis for annihilation.

Stein was fascinated by the fact that this manuscript passed through four generations of Sa’adi’s family, traveling from Salonica to Paris, from there to Rio de Janeiro and, finally, to Jerusalem.

“It somehow eluded destruction or disappearance despite the collapse of the Salonican Jewish community and the dispersal of the author’s descendants over multiple countries and continents,” she said. “It knit together a family even as the historic Sephardi heartland of southeastern Europe was unraveling.”

Stein’s book traces decades of family correspondence and shared memories to reveal what became of Sa’adi’s 14 children and their far-flung descendants. Most fled Salonica after the collapse of the Ottoman Empire, or attempted to flee later in 20th Century, when 37 members of the Levy family perished in the Holocaust.

This article originally appeared in the UCLA Newsroom.

L.A. could better target homeless prevention services with predictive analytics

Photograph of homeless tent encampment.

Photo credit: California Policy Lab

Each year, 2 million single adults receive housing, health, and emergency services from Los Angeles County. About 2% of them — around 76,000 people — will become homeless. Predictive modeling could help address the homelessness crisis in Los Angeles County, according to a report by researchers from the California Policy Lab at UCLA, and the Poverty Lab at the University of Chicago Harris School of Public Policy.

Using data from seven Los Angeles County agencies about services they provided to county residents between 2012 and 2016 — the residents’ names and personally identifiable information were omitted and each person was assigned an ID number for the study — researchers developed a model to predict which 3,000 residents were most likely to become homeless in 2017.

The researchers then checked the accuracy of their predictions against county records, and found that 46% of the people predicted by the model to be at risk for first-time homelessness or a repeat spell of homelessness did in fact become homeless at some point during 2017.

“Bringing together data from multiple county agencies gave us a more nuanced understanding about what’s happening to people right before they slip into homeless and how services can be better targeted to prevent that from happening,” said Till von Wachter, a UCLA economics professor and co-author of the report.  Von Wachter is also faculty director at the California Policy Lab.

The California Policy Lab pairs UCLA and other UC researchers with policymakers to solve urgent social problems, including homelessness, poverty, crime and education inequality.

The research informed an action plan that was developed by the county-led Mainstream Systems Homelessness Prevention Workgroup. That plan, which was submitted to the Los Angeles County Board of Supervisors on Dec. 16, recommends that the county use predictive models to intervene with adults who are identified as having a high risk for homelessness before they reach a crisis.

It also suggests that the county launch a multidisciplinary homelessness prevention unit that includes representatives of the county’s departments of mental health, health services and social services, and the sheriff and probation offices. The unit would accept referrals from the risks lists generated by the predictive models, identify which programs or services would be most helpful for each individual, and then reach out to people to connect them to those services.

The plan is expected to receive $3 million in funding during 2020 from Measure H, a sales tax approved by Los Angeles County voters in 2017 to help address the homeless crisis, in addition to drawing some existing resources from Los Angeles County departments.

“Last year, despite providing housing to tens of thousands of people, we saw more and more individuals and families becoming homeless,” said Phil Ansell, the director of the Los Angeles County Homeless Initiative. “The county is focused on using strategic approaches to preventing homelessness, and these groundbreaking models will make it possible to reach those who need us the most before they reach the crisis point and fall into homelessness.”

The models allowed researches to identify warning signs that could help local governments intervene early, especially for residents living in deep poverty, said Harold Pollack, the Helen Ross Professor at the University of Chicago School of Social Service Administration and a co-author of the study. The research that led to the new recommendations was begun at Chicago.

“The models suggest that sharp spikes in service use, increasingly frequent service use and the receipt of multiple services from a single agency are all warning signs that someone is at high risk for homelessness,” Pollack said. “We’re now diving deeper into the models with our Los Angeles County partners to learn more and to see how these results can help focus public health and social services to this vulnerable population.”

Janey Rountree, executive director of the California Policy Lab at UCLA, said using the predictive models could go a long way toward making sure homeless prevention services reach the right people at the right time.

“Predictive modeling can help ensure that happens, before they’re in a full-blown crisis,” she said. “We look forward to seeing its impact in connecting people to the help they need.”

The study also found:

  • Effectively serving the 1% of county clients who have the greatest risk for a new homeless spell would prevent nearly 6,900 homeless spells in one year.
  • County residents who have the highest risk for homelessness are interacting with multiple agencies.
  • Falling into homelessness happens very quickly, typically within six months of a precipitating event, meaning that Los Angeles County and service providers must react quickly.

The research was provided at no cost to the county. Financial support was provided by Arnold Ventures and the Max Factor Family Foundation.

This article originally appeared in the UCLA Newsroom.

Chronic opioid treatment may raise risk of post-traumatic stress disorder, study finds

Senior author Michael Fanselow said the research suggests that chronic opioid use increases susceptibility to developing anxiety disorders. Photo credit: Reed Hutchinson/UCLA

While opioids are often prescribed to treat people with trauma-related pain, a new UCLA-led study suggests doctors should use caution before prescribing the drug to those they believe may experience severe stress in the future, in order to reduce the risk the patient will develop PTSD.

In the study, researchers administered doses of the opioid morphine to a group of 22 mice for one week, then gave the mice relatively strong foot shocks. After the morphine wore off, the mice were given mild electric foot shocks. These mice showed a substantially longer “freezing response” than a second, control group of 24 mice that had not been given morphine. When mice recall a frightening memory, they freeze. Their heart rates and blood pressure go up, and the more frightening the memory, the more they freeze.

“While we are generally aware that drug use, such as that in the current opioid crisis, has many deleterious effects, our results suggest yet another effect — increased susceptibility to developing anxiety disorders,” said senior author Michael Fanselow, UCLA distinguished Staglin family professor of psychology and director of UCLA’s Staglin Family Music Festival Center for Brain and Behavioral Health. “As opioids are often prescribed to treat symptoms such as pain that may accompany trauma, caution may be needed because this may lead to a greater risk of developing PTSD, if exposed to further traumatic events, such as an accident, later on.”

“The foot shocks produced lasting fear and anxiety-like behaviors, such as freezing,” Fanselow said.

“Our data are the first to show a possible effect of opioids on future fear learning, suggesting that a person with a history of opioid use may become more susceptible to the negative effects of stress,” Fanselow said. “The ability of opioids to increase PTSD-like symptoms far outlasted the direct effects of the drug or withdrawal from the drug, suggesting the effect may continue even after opioid treatment has stopped.”

Fanselow’s view is if there is reason to believe a patient is likely to experience severe emotional stress after opioid treatment, then doctors should use caution about prescribing an opioid. If opioid use is medically called for, then the patient should be kept away from potentially stressful situations. So, for example, a soldier treated with opioids for pain should not be sent back into combat for a period of time, he said. The development of post-traumatic stress disorder requires some stressful experience after opioid use, he said.

The researchers also gave some of the mice morphine after the initial trauma had occurred but before exposing them to the second, mild stressor. They found that mice treated with morphine after the initial trauma did not show enhanced fear learning following exposure to the mild stressor. This finding suggests that chronic use of opioids before — but not after — a traumatic event occurs affects fear learning during subsequent stressful events.

The researchers concluded the mice given morphine were more susceptible to post-traumatic stress disorder than the control group of mice not given any opioids, and inferred that people with a history of using opioids are more susceptible to PTSD than the general population.

The study is published in Neuropsychopharmacology, an international scientific journal focusing on clinical and basic science research that advances understanding of the brain and behavior.

The research was funded by the National Institute on Drug Abuse and National Institute of Mental Health.

An opiate is a drug naturally derived from the opium poppy plant, such as heroin, morphine and codeine. Opioid is a broader term that includes opiates and any substance, natural or synthetic, that binds to the brain’s opioid receptors — which play a key role in controlling pain, rewards and addictive behaviors. Synthetic opioids include the prescription painkillers Vicodin and OxyContin, as well as fentanyl and methadone.

Substance abuse and PTSD often go hand-in-hand, Fanselow said, and people with PTSD often take drugs to self-medicate. Nearly 40% of people with PTSD also have a drug disorder.

Fanselow and colleagues reported last month that a traumatic brain injury causes changes in a brain region called the amygdala; and the brain processes fear differently after such an injury.

This article originally appeared in the UCLA Newsroom.

New NASA image provides more details about first observed interstellar comet

Image of interstellar comet.

The interstellar comet Comet 2I/Borisov (blueish image at right) near a spiral galaxy (left), in an image taken Nov. 16. Photo credit: NASA, ESA and David Jewitt/UCLA

A new image from NASA’s Hubble Space Telescope provides important new details about the first interstellar comet astronomers have seen in our solar system.

The comet, called Comet 2I/Borisov (the “I” stands for interstellar), was spotted near a spiral galaxy known as 2MASX J10500165-0152029. It was approximately 203 million miles from Earth when the image was taken on Nov. 16.

“Data from the Hubble Space Telescope give us the best measure of the size of comet 2I/Borisov’s nucleus, which is the really important part of the comet,” said David Jewitt, a UCLA professor of planetary science and astronomy who analyzed and interpreted the data from the new image.

Jewitt collaborated on the new analysis with colleagues from the University of Hawaii, Germany’s Max Planck Institute for Solar System Research, the Space Telescope Science Institute in Baltimore and Johns Hopkins University’s Applied Physics Laboratory. The scientists were surprised to learn that the nucleus has a radius measuring only about half of a kilometer — or less than one-fifteenth the size that earlier investigations suggested it might be.

“That is important because knowing its size helps us to determine the total number, and mass, of other similar objects in the solar system and the Milky Way,” Jewitt said. “2I/Borisov is the first known interstellar comet, and we would like to learn how many others there are.”

The comet is traveling at a breathtaking speed of 110,000 miles per hour — one of the fastest comets ever seen, Jewitt said. More commonly, comets travel at about half that speed.

Crimean astronomer Gennady Borisov discovered the comet on Aug. 30, using a telescope he built. Based on precise measurements of its changing position, the International Astronomical Union’s Minor Planet Center calculated a likely orbit for the comet, which shows that it came from elsewhere in the galaxy. Jewitt said its precise point of origin is unknown.

A second Hubble Space Telescope image of the comet, taken on Dec. 9, shows the comet even closer to Earth, approximately 185 million miles from Earth, he said.

Comets are icy bodies thought to be fragments left behind when planets form in the outer parts of planetary systems.

Observations by numerous telescopes show that the comet’s chemical composition is similar to that of comets previously observed in our solar system, which provides evidence that comets also form around other stars, Jewitt said. By mid-2020, the comet will have zoomed past Jupiter on its way back into interstellar space, where it will drift for billions of years, Jewitt said.

This article originally appeared in the UCLA Newsroom.

One of California’s iconic tree species offers lessons for conservation

Picture of a valley oak tree.

The valley oak, the largest oak in California, grows to over 100 feet tall and provides habitat and food for a variety of animals. Photo credit: Victoria Sork/UCLA

 

With increasing regularity, Californians are witnessing firsthand the destructive power of wildfires. But not everyone sees what happens after the flames die down, when debris is cleared, homes and lives rebuilt — and trees replanted to help nature recover.

New research led by UCLA evolutionary biologist Victoria Sork examines whether the trees being replanted in the wake of California’s fires will be able to survive a climate that is continuing to warm.

The study, which is published in the Proceedings of the Natural Academy of Sciences, focuses on California’s iconic valley oak. The research is among the first to demonstrate the potential of using genomics to inform conservation strategies — essentially giving species an evolutionary boost. The study showed that planting trees that are genetically better suited to higher temperatures makes them more likely to survive and grow to maturity.

“When we think about managing ecosystems under rapidly changing climates, we have to realize trees need to be able to survive past 50 years,” Sork said.

The paper also discovered something surprising: The valley oak, an essential component of many ecosystems in California, is already poorly adapted to its environment — even considering climate conditions in 2019.

“They actually seem to grow better in cooler climates than they’re in right now,” said Luke Browne, a postdoctoral scholar at the UCLA La Kretz Center for California Conservation Science and the study’s lead author. “They might grow better if climates were more like they were 21,000 years ago, during the last ice age.”

During the peak of the last ice age, summer temperatures were about 4 to 5 degrees Celsius colder, and ice covered most of Canada and mountainous areas of the U.S.

In the fields of conservation and land management, it is a common assumption that plants and animals are adapted to their environments — that’s how evolution and natural selection are supposed to work. The new research casts doubt on that assumption.

The study is part of an ongoing project initiated by Sork and Jessica Wright, an expert in conservation genetics at the USDA Forest Service, more than 10 years ago.

Researchers gathered 11,000 seeds from 94 locations throughout the trees’ range, which stretches from the Santa Monica Mountains to the Cascade foothills in the northern part of the state. They grew them to saplings in a greenhouse and planted them in two large experimental gardens, in Chico and Placerville, California. They tracked how well trees from different locations grew, and sequenced the genomes of their mother trees to link genetic information and growth rates.

The researchers then identified which genetic variants would be more likely to thrive as climate change continues to warm California. They predicted that, under predicted future warmer temperatures, trees containing beneficial genetic variations would have 11% higher growth rates than the average for all of the trees in the experiment, and 25% higher growth rates than the trees without the beneficial variations.

Information like that could help the U.S. Forest Service, for example, in its efforts to restore forests with species that have the best chance for long-term survival.

“Studies like this one provide valuable insights that help land managers make informed decisions on reforestation projects,” Wright said. “When planting trees in a particular location, managers have to decide where to collect the acorns.”

By 2070, average temperatures in the state are projected to be up to 4.8 degrees warmer than they were during the mid- to late 20th century.

“That’s going to have consequences for how fast these trees grow,” Browne said. “We’re at a challenging time to figure out the best way to do conservation science. This paper shows one approach we could use that takes advantage of modern genomics.”

The study did not determine why valley oaks are not well adapted to their environment. It might be because the climate has already warmed up so much, the trees’ long lifespans — up to 500 years — or some other, unknown factor.

The valley oak is the largest oak in California; it grows to over 100 feet tall, and has dark green leaves and a deeply grooved trunk. It is considered a foundational species because it provides habitat and food for a variety of animals, including squirrels, birds, deer and insects. In parts of the state, it is one of the only species of tree that exists. Valley oaks provide benefits to humans, too: filtering water and providing shady places to escape the heat.

Although it focuses on the oak, the paper has broader implications for conservation science in a changing climate — especially for species that evolve and adapt slowly. That’s what Sork and Wright were thinking when they initiated the project.

At the time, they hoped to find conservation strategies that could eventually be implemented using genetic information alone — without extensive field experiments.

“Not everyone in the world is going to be able to collect 11,000 seeds and plant them in a common garden,” Sork said.

This article originally appeared in the UCLA Newsroom.

Black hole at the center of our galaxy appears to be getting hungrier

Rendering of a star called S0-2 orbiting the supermassive black hole at the center of the Milky Way. It did not fall in, but its close approach could be one reason for the black hole’s growing appetite. Photo credit: Nicolle Fuller/National Science Foundation

The enormous black hole at the center of our galaxy is having an unusually large meal of interstellar gas and dust, and researchers don’t yet understand why.

“We have never seen anything like this in the 24 years we have studied the supermassive black hole,” said Andrea Ghez, UCLA professor of physics and astronomy and a co-senior author of the research. “It’s usually a pretty quiet, wimpy black hole on a diet. We don’t know what is driving this big feast.”

paper about the study, led by the UCLA Galactic Center Group, which Ghez heads, is published today in Astrophysical Journal Letters.

The researchers analyzed more than 13,000 observations of the black hole from 133 nights since 2003. The images were gathered by the W.M. Keck Observatory in Hawaii and the European Southern Observatory’s Very Large Telescope in Chile. The team found that on May 13, the area just outside the black hole’s “point of no return” (so called because once matter enters, it can never escape) was twice as bright as the next-brightest observation.

They also observed large changes on two other nights this year; all three of those changes were “unprecedented,” Ghez said.

The brightness the scientists observed is caused by radiation from gas and dust falling into the black hole; the findings prompted them to ask whether this was an extraordinary singular event or a precursor to significantly increased activity.

“The big question is whether the black hole is entering a new phase — for example if the spigot has been turned up and the rate of gas falling down the black hole ‘drain’ has increased for an extended period — or whether we have just seen the fireworks from a few unusual blobs of gas falling in,” said Mark Morris, UCLA professor of physics and astronomy and the paper’s co-senior author.

The team has continued to observe the area and will try to settle that question based on what they see from new images.

“We want to know how black holes grow and affect the evolution of galaxies and the universe,” said Ghez, UCLA’s Lauren B. Leichtman and Arthur E. Levine Professor of Astrophysics. “We want to know why the supermassive hole gets brighter and how it gets brighter.”

► UCLA astronomers discussed the project in a Keck Observatory video

The new findings are based on observations of the black hole — which is called Sagittarius A*, or Sgr A* — during four nights in April and May at the Keck Observatory. The brightness surrounding the black hole always varies somewhat, but the scientists were stunned by the extreme variations in brightness during that timeframe, including their observations on May 13.

“The first image I saw that night, the black hole was so bright I initially mistook it for the star S0-2, because I had never seen Sagittarius A* that bright,” said UCLA research scientist Tuan Do, the study’s lead author. “But it quickly became clear the source had to be the black hole, which was really exciting.”

One hypothesis about the increased activity is that when a star called S0-2 made its closest approach to the black hole during the summer 2018, it launched a large quantity of gas that reached the black hole this year.

Another possibility involves a bizarre object known as G2, which is most likely a pair of binary stars, which made its closest approach to the black hole in 2014. It’s possible the black hole could have stripped off the outer layer of G2, Ghez said, which could help explain the increased brightness just outside the black hole.

Morris said another possibility is that the brightening corresponds to the demise of large asteroids that have been drawn in to the black hole.

No danger to Earth

The black hole is some 26,000 light-years away and poses no danger to our planet. Do said the radiation would have to be 10 billion times as bright as what the astronomers detected to affect life on Earth.

Astrophysical Journal Letters also published a second article by the researchers, describing speckle holography, the technique that enabled them to extract and use very faint information from 24 years of data they recorded from near the black hole.

Ghez’s research team reported July 25 in the journal Science the most comprehensive test of Einstein’s iconic general theory of relativity near the black hole. Their conclusion that Einstein’s theory passed the test and is correct, at least for now, was based on their study of S0-2 as it made a complete orbit around the black hole.

► Watch a four-minute film about Ghez’s research

Ghez’s team studies more than 3,000 stars that orbit the supermassive black hole. Since 2004, the scientists have used a powerful technology that Ghez helped pioneer, called adaptive optics, which corrects the distorting effects of the Earth’s atmosphere in real time. But speckle holography enabled the researchers to improve the data from the decade before adaptive optics came into play. Reanalyzing data from those years helped the team conclude that they had not seen that level of brightness near the black hole in 24 years.

“It was like doing LASIK surgery on our early images,” Ghez said. “We collected the data to answer one question and serendipitously unveiled other exciting scientific discoveries that we didn’t anticipate.”

Co-authors include Gunther Witzel, a former UCLA research scientist currently at Germany’s Max Planck Institute for Radio Astronomy; Mark Morris, UCLA professor of physics and astronomy; Eric Becklin, UCLA professor emeritus of physics and astronomy; Rainer Schoedel, a researcher at Spain’s Instituto de Astrofısica de Andalucıa; and UCLA graduate students Zhuo Chen and Abhimat Gautam.

The research is funded by the National Science Foundation, W.M. Keck Foundation, the Gordon and Betty Moore Foundation, the Heising-Simons Foundation, Lauren Leichtman and Arthur Levine, and Howard and Astrid Preston.

This article originally appeared in the UCLA Newsroom.

UCLA historian Kelly Lytle Hernández awarded MacArthur Fellowship

Kelly Lytle Hernández, a 2019 MacArthur Foundation Fellow, is one of 14 UCLA faculty to be chosen for the honor. Photo credit: John D. & Catherine T. MacArthur Foundation

UCLA professor Kelly Lytle Hernández, an award-winning author and scholar of race, mass incarceration and immigration, was announced today as a recipient of a prestigious MacArthur Fellowship from the John D. and Catherine T. MacArthur Foundation.

Lytle Hernández, who is a professor of history and African American studies, is the director of UCLA’s Ralph J. Bunche Center for African American Studies, which under her leadership has focused on supporting research into two critical themes in the modern black world — work and justice. The Bunche Center is home to Million Dollar Hoods, which maps the fiscal and human cost of mass incarceration in Los Angeles. Lytle Hernández is the director and principal investigator on the project.

“Lytle Hernández’s investigation of the intersecting histories of race, mass incarceration, immigration, and cross-border politics is deepening our understanding of how imprisonment has been used as a mechanism for social control in the United States,” the foundation said.

The MacArthur Fellowship is a $625,000, no-strings-attached award to people the foundation deems “extraordinarily talented and creative individuals.” Fellows are chosen based on three criteria: exceptional creativity, promise for important future advances based on a track record of accomplishments, and potential for the fellowship to facilitate subsequent creative work. Lytle Hernández is one of 26 individuals the foundation selected for fellowships in 2019.

“As a scholar, I both work deeply alone and deeply in community, but until very recently the scholarly communities I’ve worked in — immigration and the carceral state — have been fairly separate,” said Lytle Hernández, who holds the Thomas E. Lifka Chair in History at UCLA. “I hope my work has helped people understand immigration as another aspect of mass incarceration in the United States and that my award further helps people understand that these two regimes are intertwined. This award will help us continue this work across communities and shine a light on this kind of thinking that unites these two crises that others often see as distinct.”

Lytle Hernández, 45, received a her bachelor’s degree from UC San Diego in 1996 and earned her doctorate in 2002 from UCLA.

For her first book, “MIGRA! A History of the U.S. Border Patrol,” Lytle Hernández pored over historical records to illuminate the border patrol’s nearly exclusive focus on policing unauthorized immigration from Mexico.

In “City of Inmates: Conquest, Rebellion, and the Rise of Human Caging in Los Angeles,” she began zeroing in on another dimension of race and law enforcement, specifically what forces shaped Los Angeles so that it came to operate the largest jail system in the United States.

“What I found in the archives is that since the very first days of U.S. rule in Los Angeles — the Tongva Basin — incarceration has persistently operated as a means of purging, removing, caging, containing, erasing, disappearing and otherwise eliminating indigenous communities and racially targeted populations,” Lytle Hernández said in an interview about the book.

The MacArthur Fellowship, which is commonly referred to as the “genius grant,” is according to the foundation, intended to encourage people of outstanding talent to pursue their own creative, intellectual and professional inclinations. Recipients may be writers, scientists, artists, social scientists, humanists, teachers, entrepreneurs, or those in other fields, with or without institutional affiliations.

Lytle Hernández joins 13 other UCLA faculty as MacArthur fellows, including mathematician Terence Tao, choreographer Kyle Abraham, director Peter Sellars, astrophysicist Andrea Ghez and historian of religion Gregory Schopen.

While unsure of her specific plans for the award, Lytle Hernández said that she will continue to expand the scope and scale of her social justice scholarship, including with partners outside of UCLA.

“I’d like to create a space for myself and others — especially community organizers and movement-driven scholars — to write,” she said, noting that these people’s calendars tend to be jammed by the “urgency of their work.” “I’d like to create space that allows myself and others to process the work that we’re doing and to share it.”

This article originally appeared in the UCLA Newsroom.

Study shows how serotonin and a popular anti-depressant affect the gut’s microbiota

Senior author Elaine Hsiao says researchers hope to build on their current study to learn whether microbial interactions with antidepressants have consequences for health and disease. Photo: Reed Hutchinson/UCLA

A new study in mice led by UCLA biologists strongly suggests that serotonin and drugs that target serotonin, such as anti-depressants, can have a major effect on the gut’s microbiota — the 100 trillion or so bacteria and other microbes that live in the human body’s intestines.

Serotonin — a neurotransmitter, or chemical messenger that sends messages among cells — serves many functions in the human body, including playing a role in emotions and happiness. An estimated 90% of the body’s serotonin is produced in the gut, where it influences gut immunity.

The team — led by senior author Elaine Hsiao and lead author Thomas Fung, a postdoctoral fellow — identified a specific gut bacterium that can detect and transport serotonin into bacterial cells. When mice were given the antidepressant fluoxetine, or Prozac, the biologists found this reduced the transport of serotonin into their cells. This bacterium, about which little is known, is called Turicibacter sanguinis. The study is published this week in the journal Nature Microbiology.

“Our previous work showed that particular gut bacteria help the gut produce serotonin. In this study, we were interested in finding out why they might do so,” said Hsiao, UCLA assistant professor of integrative biology and physiology, and of microbiology, immunology and molecular genetics in the UCLA College; and of digestive diseases in the David Geffen School of Medicine at UCLA.

Hsiao and her research group reported in the journal Cell in 2015 that in mice, a specific mixture of bacteria, consisting mainly of Turicibacter sanguinis and Clostridia, produces molecules that signal to gut cells to increase production of serotonin. When Hsiao’s team raised mice without the bacteria, more than 50% of their gut serotonin was missing. The researchers then added the bacteria mixture of mainly Turicibacter and Clostridia, and their serotonin increased to a normal level.

That study got the team wondering why bacteria signal to our gut cells to make serotonin. Do microbes use serotonin, and if so, for what?

In this new study, the researchers added serotonin to the drinking water of some mice and raised others with a mutation (created by altering a specific serotonin transporter gene) that increased the levels of serotonin in their guts. After studying the microbiota of the mice, the researchers discovered that the bacteria Turicibacter and Clostridia increased significantly when there was more serotonin in the gut.

If these bacteria increase in the presence of serotonin, perhaps they have some cellular machinery to detect serotonin, the researchers speculated. Together with study co-author Lucy Forrest and her team at the National Institutes of Health’s National Institute of Neurological Disorders and Stroke, the researchers found a protein in multiple species of Turicibacter that has some structural similarity to a protein that transports serotonin in mammals. When they grew Turicibacter sanguinis in the lab, they found that the bacterium imports serotonin into the cell.

In another experiment, the researchers added the antidepressant fluoxetine, which normally blocks the mammalian serotonin transporter, to a tube containing Turicibacter sanguinisThey found the bacterium transported significantly less serotonin.

The team found that exposing Turicibacter sanguinis to serotonin or fluoxetine influenced how well the bacterium could thrive in the gastrointestinal tract. In the presence of serotonin, the bacterium grew to high levels in mice, but when exposed to fluoxetine, the bacterium grew to only low levels in mice.

“Previous studies from our lab and others showed that specific bacteria promote serotonin levels in the gut,” Fung said. “Our new study tells us that certain gut bacteria can respond to serotonin and drugs that influence serotonin, like anti-depressants. This is a unique form of communication between bacteria and our own cells through molecules traditionally recognized as neurotransmitters.”

The team’s research on Turicibacter aligns with a growing number of studies reporting that anti-depressants can alter the gut microbiota. “For the future,” Hsiao said, “we want to learn whether microbial interactions with antidepressants have consequences for health and disease.” Hsiao wrote a blog post for the journal about the new research.

Other study co-authors are Helen Vuong, Geoffrey Pronovost, Cristopher Luna, Anastasia Vavilina, Julianne McGinn and Tomiko Rendon, all of UCLA; and Antoniya Aleksandrova and Noah Riley, members of Forrest’s team.

The research was supported by funding from the National Institutes of Health’s Director’s Early Independence Award, Klingenstein-Simons Fellowship Award, and David & Lucile Packard Foundation’s Packard Fellowship for Science and Engineering.

This article originally appeared in the UCLA Newsroom.

Technique could make better membranes for next-generation filtration

Photo of researchers in lab.

UCLA doctoral student Mackenzie Anderson, postdoctoral scholar Brian McVerry and professor Richard Kaner. Photo Credit: Marc Roseboro/UCLA

Deriving drinkable water from seawater, treating wastewater and conducting kidney dialysis are just a few important processes that use a technology called membrane filtration.

The key to the process is the membrane filter — a thin, semi-porous film that allows certain substances such as water to pass through while separating out other, unwanted substances. But in the past 30 years, there have been no significant improvements in the materials that make up the key layers of commercially produced membrane filters.

Now, UCLA researchers have developed a new technique called thin-film liftoff, or T-FLO, for creating membrane filters. The approach could offer a way for manufacturers to produce more effective and energy-efficient membranes using high-performance plastics, metal-organic frameworks and carbon materials. To date, limitations in how filters are fabricated have prevented those materials from being viable in industrial production.

A study describing the work is published in the journal Nano Letters.

“There are a lot of materials out there that in the lab can do nice separations, but they’re not scalable,” said Richard Kaner, UCLA’s Dr. Myung Ki Hong Professor of Materials Innovation and the study’s senior author. “With this technique, we can take these materials, make thin films that are scalable, and make them useful.”

In addition to their potential for improving types of filtration that are performed using current technology, membranes produced using T-FLO could make possible an array of new forms of filtration, said Kaner, who also is a distinguished professor of chemistry and biochemistry, and of materials science and engineering, and a member of the California NanoSystems Institute at UCLA. For example, the technique might one day make it feasible to pull carbon dioxide out of industrial emissions — which would enable the carbon to be converted to fuel or other applications while also reducing pollution.

Filters like the ones used for desalination are called asymmetric membranes because of their two layers: a thin but dense “active” layer that rejects particles larger than a specific size, and a porous “support” layer that gives the membrane structure and allows it to resist the high pressures used in reverse osmosis and other filtering processes. The first asymmetric membrane for desalination was devised by UCLA engineers in the 1960s.

Today’s asymmetric membranes are made by casting the active layer onto the support layer, or casting both concurrently. But to manufacture an active layer using more advanced materials, engineers have to use solvents or high heat — both of which damage the support layer or prevent the active layer from adhering.

In the T-FLO technique, the active layer is cast as a liquid on a sheet of glass or metal and cured to make the active layer solid. Next, a support layer made of epoxy reinforced with fabric is added and the membrane is heated to solidify the epoxy.

The use of epoxy in the support layer is the innovation that distinguishes the T-FLO technique — it enables the active layer to be created first so that it can be treated with chemicals or high heat without damaging the support layer.

The membrane then is submerged in water to wash out the chemicals that induce pores in the epoxy and to loosen the membrane from the glass or metal sheet.

Finally, the membrane is peeled off of the plate with a blade — the “liftoff” that gives the method its name.

“Researchers around the world have demonstrated many new exciting materials that can separate salts, gases and organic materials more effectively than is done industrially,” said Brian McVerry, a UCLA postdoctoral scholar who invented the T-FLO process and is the study’s co-first author. “However, these materials are often made in relatively thick films that perform the separations too slowly or in small samples that are difficult to scale industrially.

“We have demonstrated a platform that we believe will enable researchers to use their new materials in a large, thin, asymmetric membrane configuration, testable in real-world applications.”

The researchers tested a membrane produced using T-FLO for removing salt from water, and it showed promise for solving one of the common problems in desalination, which is that microbes and other organic material can clog the membranes. Although adding chlorine to the water can kill the microbes, the chemical also causes most membranes to break down. In the study, the T-FLO membrane both rejected the salt and resisted the chlorine.

In other experiments, the new membrane was also able to remove organic materials from solvent waste and to separate greenhouse gases.

Mackenzie Anderson, a UCLA doctoral student, is co-first author of the study.

The research was supported by the U.S./China Clean Energy Research Center for Water-Energy Technologies and the National Science Foundation. The project is aligned with UCLA’s Sustainable LA Grand Challenge.

Among the many other devices developed by Kaner’s laboratory is a commercial membrane that separates oil from water and cleans up the debris left by fracking. Fracking is a technique that uses high-pressure mixes of water, sand or gravel and chemicals to extract gas and oil from shale rock.

Kaner is among the world’s most highly cited scientific researchers, and he was one of the 2019 recipients of the American Institute of Chemists’ Chemical Pioneer Award, which honors chemists and chemical engineers for contributions that advance the science of chemistry or the chemical profession.

This article originally appeared in the UCLA Newsroom.

Biochemists discover new insights into what may go awry in brains of people with Alzheimer’s

Photo of two researchers in lab.

Research by UCLA professor Steven Clarke and former graduate student Rebeccah Warmack, as well as UCLA colleagues, reveals new information about the brain’s biochemistry.

More than three decades of research on Alzheimer’s disease have not produced any major treatment advances for those with the disorder, according to a UCLA expert who has studied the biochemistry of the brain and Alzheimer’s for nearly 30 years. “Nothing has worked,” said Steven Clarke, a distinguished professor of chemistry and biochemistry. “We’re ready for new ideas.” Now, Clarke and UCLA colleagues have reported new insights that may lead to progress in fighting the devastating disease.

Scientists have known for years that amyloid fibrils — harmful, elongated, water-tight rope-like structures — form in the brains of people with Alzheimer’s, and likely hold important clues to the disease. UCLA Professor David Eisenberg and an international team of chemists and molecular biologists reported in the journal Nature in 2005 that amyloid fibrils contain proteins that interlock like the teeth of a zipper. The researchers also reported their hypothesis that this dry molecular zipper is in the fibrils that form in Alzheimer’s disease, as well as in Parkinson’s disease and two dozen other degenerative diseases. Their hypothesis has been supported by recent studies.

Alzheimer’s disease, the most common cause of dementia among older adults, is an irreversible, progressive brain disorder that kills brain cells, gradually destroys memory and eventually affects thinking, behavior and the ability to carry out the daily tasks of life. More than 5.5 million Americans, most of whom are over 65, are thought to have dementia caused by Alzheimer’s.

The UCLA team reports in the journal Nature Communications that the small protein beta amyloid, also known as a peptide, that plays an important role in Alzheimer’s has a normal version that may be less harmful than previously thought and an age-damaged version that is more harmful.

Rebeccah Warmack, who was a UCLA graduate student at the time of the study and is its lead author, discovered that a specific version of age-modified beta amyloid contains a second molecular zipper not previously known to exist. Proteins live in water, but all the water gets pushed out as the fibril is sealed and zipped up. Warmack worked closely with UCLA graduate students David Boyer, Chih-Te Zee and Logan Richards; as well as senior research scientists Michael Sawaya and Duilio Cascio.

What goes wrong with beta amyloid, whose most common forms have 40 or 42 amino acids that are connected like a string of beads on a necklace?

The researchers report that with age, the 23rd amino acid can spontaneously form a kink, similar to one in a garden hose. This kinked form is known as isoAsp23. The normal version does not create the stronger second molecular zipper, but the kinked form does.

“Now we know a second water-free zipper can form, and is extremely difficult to pry apart,” Warmack said. “We don’t know how to break the zipper.”

The normal form of beta amyloid has six water molecules that prevent the formation of a tight zipper, but the kink ejects these water molecules, allowing the zipper to form.

When one of its amino acids forms a kink, beta amyloid creates a harmful molecular zipper, shown here in green. Photo credit: Rebeccah Warmack/UCLA

When one of its amino acids forms a kink, beta amyloid creates a harmful molecular zipper, shown here in green.
“Rebeccah has shown this kink leads to faster growth of the fibrils that have been linked to Alzheimer’s disease,” said Clarke, who has conducted research on biochemistry of the brain and Alzheimer’s disease since 1990. “This second molecular zipper is double trouble. Once it’s zipped, it’s zipped, and once the formation of fibrils starts, it looks like you can’t stop it. The kinked form initiates a dangerous cascade of events that we believe can result in Alzheimer’s disease.”

Why does beta amyloid’s 23rd amino acid sometimes form this dangerous kink?

Clarke thinks the kinks in this amino acid form throughout our lives, but we have a protein repair enzyme that fixes them.

“As we get older, maybe the repair enzyme misses the repair once or twice,” he said. “The repair enzyme might be 99.9% effective, but over 60 years or more, the kinks eventually build up. If not repaired or if degraded in time, the kink can spread to virtually every neuron and can do tremendous damage.”

“The good news is that knowing what the problem is, we can think about ways to solve it,” he added. “This kinked amino acid is where we want to look.”

The research offers clues to pharmaceutical companies, which could develop ways to prevent formation of the kink or get the repair enzyme to work better; or by designing a cap that would prevent fibrils from growing.

Clarke said beta amyloid and a much larger protein tau — with more than 750 amino acids — make a devastating one-two punch that forms fibrils and spreads them to many neurons throughout the brain. All humans have both beta amyloid and tau. Researchers say it appears that beta amyloid produces fibrils that can lead to tau aggregates, which can spread the toxicity to other brain cells. However, exactly how beta amyloid and tau work together to kill neurons is not yet known.

In this study, Warmack produced crystals, both the normal and kinked types, in 15 of beta amyloid’s amino acids. She used a modified type of cryo-electron microscopy to analyze the crystals. Cryo-electron microscopy, whose development won its creators the 2017 Nobel Prize in chemistry, enables scientists to see large biomolecules in extraordinary detail. Professor Tamir Gonen pioneered the modified microscopy, called microcrystal electron diffraction, which enables scientists to study biomolecules of any size.

Eisenberg is UCLA’s Paul D. Boyer Professor of Molecular Biology and a Howard Hughes Medical Institute investigator. Other researchers are co-author Gonen, a professor of biological chemistry and physiology at the UCLA David Geffen School of Medicine and a Howard Hughes Medical Institute investigator; and Jose Rodriguez, assistant professor of chemistry and biochemistry who holds the Howard Reiss Career Development Chair.

The research was funded by the National Science Foundation, National Institutes of Health, Howard Hughes Medical Institute, and the UCLA Longevity Center’s Elizabeth and Thomas Plott Chair in Gerontology, which Clarke held for five years.

This article originally appeared in the UCLA Newsroom.