MIT Research News' Journal
 
[Most Recent Entries] [Calendar View]

Friday, February 1st, 2019

    Time Event
    12:00a
    Putting neural networks under the microscope

    Researchers from MIT and the Qatar Computing Research Institute (QCRI) are putting the machine-learning systems known as neural networks under the microscope.

    In a study that sheds light on how these systems manage to translate text from one language to another, the researchers developed a method that pinpoints individual nodes, or “neurons,” in the networks that capture specific linguistic features.

    Neural networks learn to perform computational tasks by processing huge sets of training data. In machine translation, a network crunches language data annotated by humans, and presumably “learns” linguistic features, such as word morphology, sentence structure, and word meaning. Given new text, these networks match these learned features from one language to another, and produce a translation.

    But, in training, these networks basically adjust internal settings and values in ways the creators can’t interpret. For machine translation, that means the creators don’t necessarily know which linguistic features the network captures.

    In a paper being presented at this week’s Association for the Advancement of Artificial Intelligence conference, the researchers describe a method that identifies which neurons are most active when classifying specific linguistic features. They also designed a toolkit for users to analyze and manipulate how their networks translate text for various purposes, such as making up for any classification biases in the training data.

    In their paper, the researchers pinpoint neurons that are used to classify, for instance, gendered words, past and present tenses, numbers at the beginning or middle of sentences, and plural and singular words. They also show how some of these tasks require many neurons, while others require only one or two.

    “Our research aims to look inside neural networks for language and see what information they learn,” says co-author Yonatan Belinkov, a postdoc in the Computer Science and Artificial Intelligence Laboratory (CSAIL). “This work is about gaining a more fine-grained understanding of neural networks and having better control of how these models behave.”

    Co-authors on the paper are: senior research scientist James Glass and undergraduate student Anthony Bau, of CSAIL; and Hassan Sajjad, Nadir Durrani, and Fahim Dalvi, of QCRI.  

    Putting a microscope on neurons

    Neural networks are structured in layers, where each layer consists of many processing nodes, each connected to nodes in layers above and below. Data are first processed in the lowest layer, which passes an output to the above layer, and so on. Each output has a different “weight” to determine how much it figures into the next layer’s computation. During training, these weights are constantly readjusted.

    Neural networks used for machine translation train on annotated language data. In training, each layer learns different “word embeddings” for one word. Word embeddings are essentially tables of several hundred numbers combined in a way that corresponds to one word and that word’s function in a sentence. Each number in the embedding is calculated by a single neuron.

    In their past work, the researchers trained a model to analyze the weighted outputs of each layer to determine how the layers classified any given embedding. They found that lower layers classified relatively simpler linguistic features — such as the structure of a particular word — and higher levels helped classify more complex features, such as how the words combine to form meaning.

    In their new work, the researchers use this approach to determine how learned word embeddings make a linguistic classification. But they also implemented a new technique, called “linguistic correlation analysis,” that trains a model to home in on the individual neurons in each word embedding that were most important in the classification.

    The new technique combines all the embeddings captured from different layers — which each contain information about the word’s final classification — into a single embedding. As the network classifies a given word, the model learns weights for every neuron that was activated during each classification process. This provides a weight to each neuron in each word embedding that fired for a specific part of the classification.

    “The idea is, if this neuron is important, there should be a high weight that’s learned,” Belinkov says. “The neurons with high weights are the ones more important to predicting the certain linguistic property. You can think of the neurons as a lot of knobs you need to turn to get the correct combination of numbers in the embedding. Some knobs are more important than others, so the technique is a way to assign importance to those knobs.”

    Neuron ablation, model manipulation

    Because each neuron is weighted, it can be ranked in order of importance. To that end, the researchers designed a toolkit, called NeuroX, that automatically ranks all neurons of a neural network according to their importance and visualizes them in a web interface.

    Users upload a network they’ve already trained, as well as new text. The app displays the text and, next to it, a list of specific neurons, each with an identification number. When a user clicks on a neuron, the text will be highlighted depending on which words and phrases the neuron activates for. From there, users can completely knock out — or “ablate” — the neurons, or modify the extent of their activation, to control how the network translates.

    The task of ablation was used to determine if the researchers’ method accurately pinpointed the correct high-ranking neurons. In their paper, the researchers used the method to show that, by ablating high ranking neurons in a network, its performance in classifying correlated linguistic features dipped significantly. Alternatively, when they ablated lower-ranking neurons, performance suffered, but not as dramatically.

    “After you get all these rankings, you want to see what happens when you kill these neurons and see how badly it affects performance,” Belinkov says. “That’s an important result proving that the neurons we find are, in fact, important to the classification process.”

    One interesting application for the method is helping limit biases in language data. Machine-translation models, such as Google Translate, may train on data with gender bias, which can be problematic for languages with gendered words. Certain professions, for instance, may be more often referred to as male, and others as female. When a network translates new text, it may only produce the learned gender for those words. In many online English-to-Spanish translations, for instance, “doctor” often translates into its masculine version, while “nurse” translates into its feminine version.

    “But we find we can trace individual neurons in charge of linguistic properties like gender,” Belinkov says. “If you’re able to trace them, maybe you can intervene somehow and influence the translation to translate these words more to the opposite gender … to remove or mitigate the bias.”

    In preliminary experiments, the researchers modified neurons in a network to change translated text from past to present tense with 67 percent accuracy. They modified to switch the gender of the words with 21 percent accuracy. “It’s still a work in progress,” Belinkov says. A next step, he adds, is improving the methodology to achieve more accurate ablation and manipulation.

    12:40p
    MIMIC Chest X-Ray database to provide researchers access to over 350,000 patient radiographs

    Computer vision, or the method of giving machines the ability to process images in an advanced way, has been given increased attention by researchers in the last several years. It is a broad term meant to encompass all the means through which images can be used to achieve medical aims. Applications range from automatically scanning photos taken on mobile phones to creating 3-D renderings that aid in patient evaluations on to developing algorithmic models for emergency room use in underserved areas.

    As access to a greater number of images is apt to provide researchers with a volume of data ideal for developing better and more robust algorithms, a collection of visuals that have been enhanced, or scrubbed of patients' identifying details and then highlighted in critical areas, can have massive potential for researchers and radiologists who rely on photographic data in their work.

    Last week, the MIT Laboratory for Computational Physiology, a part of the Institute for Medical Engineering and Science (IMES) led by Professor Roger Mark, launched a preview of their MIMIC-Chest X-Ray Database (MIMIC-CXR), a repository of more than 350,000 detailed chest X-rays gathered over five years from the Beth Israel Deaconess Medical Center in Boston. The project, like the lab’s previous MIMIC-III, which houses critical care patient data from over 40,000 intensive care unit stays, is free and open to academic, clinical, and industrial investigators via the research resource PhysioNet. It represents the largest selection of publicly available chest radiographs to date.

    With access to the MIMIC-CXR, funded by Philips Research, registered users and their cohorts can more easily develop algorithms for fourteen of the most common findings from a chest X-ray, including pneumonia, cardiomegaly (enlarged heart), edema (excess fluid), and a punctured lung. By way of linking visual markers to specific diagnoses, machines can readily help clinicians draw more accurate conclusions faster and thus, handle more cases in a shorter amount of time. These algorithms could prove especially beneficial for doctors working in underfunded and understaffed hospitals.

    “Rural areas typically have no radiologists,” says Research Scientist Alistair E. W. Johnson, co-developer of the database along with Tom J. Pollard, Nathaniel R. Greenbaum, and Matthew P. Lungren; Seth J. Berkowitz, director of radiology informatics innovation; Chih-ying Deng of Harvard Medical School; and Steven Horng, associate director of emergency medicine informatics at Beth Israel. “If you have a room full of ill patients and no time to consult an expert radiologist, that’s somewhere where a model can help.”

    In the future, the lab hopes to link the X-ray archive to the MIMIC-III, thus forming a database that includes both patient ICU data and images. There are currently over 9,000 registered MIMIC-III users accessing critical care data, and the MIMIC-CXR would be a boon for those in critical care medicine looking to supplement clinical data with images.

    Another asset of the database lies in its timing. Researchers at the Stanford Machine Learning Group and the Stanford Center for Artificial Intelligence in Medicine and Imaging released a similar dataset in January, collected over 15 years at Stanford Hospital. The MIT Laboratory for Computational Physiology and Stanford University groups collaborated to ensure that both datasets released could be used with minimal legwork for the interested researcher.

    “With single center studies, you’re never sure if what you’ve found is true of everyone, or a consequence of the type of patients the hospital sees, or the way it gives its care,” Johnson says. “That’s why multicenter trials are so powerful. By working with Stanford, we’ve essentially empowered researchers around the world to run their own multicenter trials without having to spend the millions of dollars that typically costs.”

    As with MIMIC-III, researchers will be able to gain access to MIMIC-CXR by first completing a training course on managing human subjects and then agreeing to cite the dataset in their published work. 

    “The next step is free text reports,” says Johnson. “We’re moving more towards having a complete history. When a radiologist is looking at a chest X-ray, they know who the person is and why they’re there. If we want to make radiologists’ lives easier, the models need to know who the person is, too.”

    1:30p
    Study evaluates China’s progress in establishing accounting measures to reinforce its Paris pledge

    The latest round of United Nations climate talks in Poland in December sought to get the world on track to meet the 2015 Paris Agreement’s long-term goal of keeping global warming well below two degrees Celsius. Toward that end, negotiators from the agreement’s nearly 200 signatory nations were asked to report on their country or region’s progress toward fulfilling its Paris pledge, or Nationally Determined Contributions (NDC). But just how accurate were those progress reports?

    That depends on the integrity of the underlying greenhouse gas emissions data that each country used to assess its performance toward meeting the emissions reduction targets spelled out in its NDC. The measurement, reporting and verification (MRV) of a country’s overall emissions and emissions reductions involves culling and validating emissions data from multiple sources, including firms — industrial, nonprofit, and government entities — in different economic sectors. Building reliable firm-based systems for emissions MRV is no easy task, especially in developing countries where misreporting of environmental data can be significant. But a new MIT-led study in Nature Climate Change identifies challenges and opportunities to achieve that goal.

    Co-authored by researchers at MIT, Tsinghua University, and Wuhan University, the study focuses on China, the world’s largest carbon dioxide (CO2) emitter. China’s climate-change mitigation strategy centers on a national emissions trading system (ETS) whose success depends upon accurate emissions reporting at the firm level.  

    Using data obtained from two of China’s pilot regional ETS programs, one in Beijing, a highly developed major city, the other in Hubei, a less developed province, the researchers compared firms’ self-reported CO2 emissions numbers with those verified by independent third parties. The average discrepancy in these numbers decreased significantly in Beijing, going from 17 percent in 2012 to 4 percent in 2014 and 2015 for approximately 400 firms. In Hubei, which launched its system one year later, the number of discrepancies started lower and showed a statistically-insignificant decrease (from 6 percent in 2014 to 5 percent in 2015).

    “We conducted multiple tests to determine if there was any evidence for manipulation or collusion in this process,” says MIT Joint Program Research Scientist Da Zhang, the lead author of the study. “While we observed no evidence of this behavior, we did find that firms increasingly reported emissions correctly over the years, resulting in fewer reporting errors and more agreement with verifiers’ numbers.”

    In Beijing, the average numbers of reporting errors per firm decreased from 3.7 to 1.9 from 2012 to 2015, with the greatest drops in errors related to inattention and misunderstanding the rules. This tracks with previous studies indicating a reduction in reporting errors after one or two survey rounds.  

    The study emphasized that building effective MRV systems at firms in China and other developing countries takes time, resources and attention to detail. Among its recommendations to increase reporting accuracy and prevent manipulation or collusion is to provide external funding from governments or multilateral entities, at least in early years, to pay the independent verifiers. If firms pay for verification, government back-checks are essential to ensure reporting integrity. The study also maintains that strong law enforcement will be necessary to punish any detected incidents of collusion between verifiers and firms.

    “Policy efforts to meet the Paris Agreement require robust monitoring, reporting and verification of greenhouse gas emissions to demonstrate real progress,” says MIT Sloan School of Management Assistant Professor Valerie Karplus, a co-author of the study and faculty affiliate of the Joint Program. “Continuously assessing performance will be important to raising confidence in the effectiveness of nascent market-based systems, especially in countries where prior experience with such systems is limited.”

    The research was supported by the National Science Foundation of China and the U.S. Energy Information Agency and a consortium of industrial sponsors and federal grants that fund the work of the MIT Joint Program.

    << Previous Day 2019/02/01
    [Calendar]
    Next Day >>

MIT Research News   About LJ.Rossia.org