We advance science and technology to benefit humanity

blue spectrum gradient blur on black background

Our Research Lens

We consider research directions through the lens of the positive impact we aspire to create with and for customers, communities, and all of society.

A researcher working in an optical lab

Foundations

Advancing human knowledge and foundational technologies.

Science and technology have never been more deeply interconnected. We aim to accelerate and democratize scientific discovery by empowering the global research community with leading-edge tools enabled by artificial intelligence, hyperscale computing and other advances in computing technology. In turn, we aim to draw upon new fundamental insights to inform and inspire new breakthroughs and computing paradigms.

AI and Microsoft Research

Cloud systems futures.

Three researchers collaborating in an office

Ensuring that technology is trustworthy and beneficial to everyone

Advances in technology can present tremendous opportunities along with equally significant risks. We aim to enable individuals and organizations to understand and govern their data and technology in alignment with human intent and in adherence with applicable regulations.

We create leading-edge tools for realizing technology ethics from principle to practice, engage at the intersection of technology and policy, and ultimately aim to ensure that the continued advancement of technology is safe, secure, and beneficial to society.

Privacy Preserving Machine Learning Innovation

Cryptography and privacy research, fate: fairness, accountability, transparency, and ethics in ai.

Two young people with low vision use the Project Torino physical programming learning set.

Empowerment

Amplifying human creativity and achievement.

We believe that technology exists to amplify human potential and creativity across the breadth of our abilities and circumstances. 

We are focused on empowering people to pursue their goals, to discover meaning in their work and play, and to stay connected to what’s important to them. We aim to create technologies that can understand our intent and context and smoothly adapt, thereby transforming the way individuals engage with machines, data, and each other.

Collaborative Intelligence

Sociotechnical systems, the new future of work.

Two researchers collaborate at a whiteboard

Empowering scientists and accelerating the discovery of new knowledge

We aim to accelerate and democratize scientific discovery by empowering the global research community with leading edge tools enabled by advanced technologies such as artificial intelligence and hyperscale computing. In turn, we aim to draw upon new fundamental insights to inform and inspire new breakthroughs in computing technology.

If we can unlock and accelerate more scientific exploration and experimentation through advances in computing, we can help achieve breakthroughs across the sciences that will positively impact our world.

Dr. Raj Jena, Oncologist, is sat in a hospital room testing the InnerEye platform

Supporting a healthy global society

We have witnessed incredible progress in advancing science and technology for improving human health, but the benefits of these advancements are unevenly distributed throughout society.

We aim to help ensure that every person on the planet can live to their physical and mental potential by advancing inclusive and equitable technologies supporting the health of individuals and populations. In pursuit of this aim, we collaborate across disciplines and institutions at the intersection of computing technology, health, and life sciences.

Microsoft Health Futures

Biomedical imaging, microsoft immunomics, biomedical nlp, microsoft genomics.

A man stands out in a field of wheat

Fostering a resilient and sustainable society

The need to address global challenges such as sustainability and pandemic response has never been more pressing.

We aim to create more robust interconnected systems of people and technology capable of identifying societal and ecological patterns and enabling coordinated action. Effective new approaches in these areas will not only help us respond to and recover from systemic disruptions but could ultimately help us anticipate and prevent them.

Societal Resilience

Environmental sustainability.

“Microsoft Research is powered by brilliant, endlessly inquisitive individuals.” —Peter Lee, CVP, Microsoft Research & Incubations

Researcher working with quantum computer.

Research culture

We are rigorous and objective:  since its founding in 1991, Microsoft Research has been committed to an academic research approach that advances our understanding of the world and how we use technology to interact with it, and with each other.

We take calculated risks:  as an industrial research lab, we contribute ideas that help improve Microsoft’s existing products and services and clear paths to new businesses – but we start from a place of curiosity and exploration. This leads us to take on challenges with an eye toward the longer-term benefit and a willingness to learn from failure as we explore new frontiers.

Ethan Jackson standing in the Microsoft Premonition display

We show and share our work:   over the past thirty years, our research community and collaborators have published tens of thousands of  papers  and released hundreds of  open-source  projects and datasets. Nearly all of the research we produce is shared with the research community.

Photo of a quantum computer close-up

We move the conversation forward: Alongside new papers and new inventions, conversations and constructive debate are key to advancing the state of the art across all the sciences and to aligning the research community with the people its research aims to serve. That’s why we share much of our research through conferences, events, and other forums where these important dialogues take place.

We engage with the entire research ecosystem: we work across disciplines – from social sciences to economics, to mathematics and physics and collaborate across the boundaries between research and engineering and with our peers in academia, industry, and government. We learn, build, and amplify research as we convene within the research ecosystem and co-create research that impacts society and the world.

Meet the people who make it possible

We work collectively to make Microsoft Research a welcoming and productive space for all researchers. Our researchers have the opportunity to create and pursue a research agenda with our group’s diverse array of researchers and practitioners, as well as contribute to ongoing research projects.

a diverse group of people brainstorming

Engage with us

a group of people standing in front of a crowd

Academic conferences

Events provide opportunities to meet our community of researchers, learn about exciting research topics, and grow your network.

Microsoft Research Forum

Join us for a continuous exchange of ideas about science and technology research in the era of general AI. This series will explore recent research advances, bold new ideas, and important discussions with the global research community.

various abstract 3D shapes on a light blue background

Academic programs

Microsoft recognizes the value of diversity in computing and aims to increase the pipeline of talent receiving advanced degrees in computing-related fields.

Research tools

We maintain an index of datasets, models, SDKs, APIs, and other open-source code created by Microsoft researchers and shared with the broader academic community.

Collaborations

Microsoft collaborates with the global research community through programs, events, learning opportunities, and joint research endeavors.

Careers in research

Join a brilliant team of researchers working to solve technology’s most exciting challenges. Come as you are. Do what you love.

  • Follow on Twitter
  • Like on Facebook
  • Follow on LinkedIn
  • Subscribe on Youtube
  • Follow on Instagram
  • Subscribe to our RSS feed

Share this page:

  • Share on Twitter
  • Share on Facebook
  • Share on LinkedIn
  • Share on Reddit

microsoft research articles

Microsoft and LinkedIn share latest data and innovation for hybrid work

Sep 9, 2021 | Jared Spataro - CVP Modern Work & Business Applications

  • Share on Facebook (opens new window)
  • Share on Twitter (opens new window)
  • Share on LinkedIn (opens new window)

Coworkers have Teams meeting remotely

This morning Satya  Nadella  and Ryan  Roslan sky , the CEO of LinkedIn,   sat down  to talk about the  key secular trends   they’re  seeing as  people and organizations  everywhere  adjust to  hybrid work .   Their conversation  is part of  a n  effort  between  Microsoft and LinkedIn to help leaders  and their teams  navigate this new world of work, and here I’ll share  more about the trends we’re seeing, the  data behind them  and  the   product updates we ’ re announcing  today across Microsoft Teams, Microsoft Viva and LinkedIn .   

As Satya and Ryan mentioned, the evolving Delta variant is compelling many of us to adjust plans for reopening worksites. It’s a stark reminder that  this is the new normal. Our ability to come together will ebb and flow.  In fact, we had planned for Oct. 4 to be the first possible date to fully reopen Microsoft’s own Redmond headquarters, and many other work sites in the U.S. But as we shared with our employees today, we’ve shifted those plans. Given the uncertainty of COVID-19, we’ve decided against attempting to forecast a new date for a full reopening of our U.S. work sites in favor of opening U.S. work sites as soon as we’re able to do so safely based on public health guidance. From there, we’ll communicate a 30-day transition period that provides time for employees to prepare while allowing us to continue to be agile and flexible as we look to the data and make choices to protect employee health, safety and well-being.  

We’ve heard many business leaders come forward with strong opinions on how,  when and where people should work in a hybrid world. At Microsoft and LinkedIn, we want to take a learn-it-all approach, and lead with data rather than dogma . And we’re incorporating flexibility into our decision-making. Because in uncharted territory, we need to be able to shift and adjust as data and research offer new insights to guide our way.   

The Hybrid Work Paradox and the ‘Great Reshuffle’  

A report out today on our  Work Trend Index shares what we’re learning from Microsoft employees in over 100 countries around the world. Employee surveys tell us that while hybrid work is complex, embracing flexibility, different work styles and a culture of trust can help us all navigate it successfully. In a year when we sent 160,000 people home to work and remotely onboarded 25,000 new employees, the share of people who report feeling included at Microsoft is at an all-time high of 90%. According to surveys, employee confidence and support from our managers is also at an all-time high. While there is still so much more we can do in our culture journey at Microsoft – and we’d love to see these numbers even higher – this data show that we don’t have to be physically together to feel like we’re in it together.  

But there’s no guarantee that these positive trends will continue in hybrid, and difficulties remain. As we navigate remote work during a pandemic, our employee surveys show continued challenges to satisfaction with work-life balance and team connection.  

While we hope hybrid work will help us improve in these areas, finding the balance will be complex. Our ongoing  research  shows employees crave more in-person time with their team but wish to keep the flexibility of remote work. And every person is different – 58% of employees who plan to spend the most and least time in-office are doing it for the same reason: more focused work. And there are gaps to fill – managers plan to spend a higher share of their time in-office than non-managerial employees (45% vs. 39%). Moreover, employees surveyed plan to go into the office more than managers expect. This complexity is what Satya calls the  Hybrid Work Paradox .  

Chart

Solving the Hybrid Work Paradox will be the challenge of the decade. We need policies and technology tuned for flexibility, but policies and technology alone are not enough. Every leader must also ask: How do I rebuild social capital and connection back to mission, culture and team? How do I help people prioritize individual well-being and work in sustainable ways? As Satya has said:  “Our new data shows there is no one-size-fits-all approach to hybrid work, as employee expectations continue to change. The only way for organizations to solve for this complexity is  to embrace flexibility across their entire operating model,  including the ways people work, the places they inhabit and how they approach business process.”  

All of this will be key to navigating this moment of change that  our colleagues at LinkedIn are calling the “Great Reshuffle.”  Leaders are rethinking their working models, cultures and company values, while at the same time, employees are rethinking not only how and where they work, but  why . At the core of it all is the start of a new, more dynamic relationship between employers and employees.  

To better understand how employers are thinking about navigating this new world of work,  LinkedIn surveyed more than 500 C-level executives  in the U.S. and U.K. Top of mind for executives is the same thing on the minds of employees – flexibility. With  87% of people saying they would prefer  to stay remote at least half the time, a majority of employers are adapting: 81% of leaders are changing their workplace policies to offer greater flexibility.   

Despite all the change, leaders feel like there are opportunities ahead – more than half (58%) are optimistic that flexibility will be good for both people and the business. Leaders have an opportunity to rewrite their playbook when it comes to hiring, skills development, engaging talent and more.   

Employers and employees have an opportunity to build new relationships grounded in shared values and a common mission. And more people will be doing work that best matches their skills and needs, leading to greater success for organizations that   engage their employees with empathy and trust .  But the data also underscores the imperative for business leaders to transform themselves to attract and retain top talent as employee expectations evolve.    

Taken together, the Hybrid Work Paradox and the Great Reshuffle are creating fundamental changes in the global labor market. Ultimately, no one can predict how things will shake out. But history has shown us that there is always opportunity in turbulent times.   

Companies that are adaptable and able to continuously innovate will have the advantage. To really compete, however, they’ll need to enable sustainable productivity for their employees. Right now, the  power has shifted from employer to employee, and people will vote with their feet.  As conditions change, this pendulum of power will continue to swing in both directions, but two constants remain: First: as Satya said above, every leader and every organization will need to create a new operating model across people, places and processes. And second: those companies with a better employee experience for all employees – from the virtual boardroom to the factory floor – will be the ones to  attract and retain better talent.   

How technology can help   

Microsoft and LinkedIn are navigating the challenges and opportunities of hybrid work just like every other organization. We see the role of technology as an enabler – helping our employees and our customers as they transform for hybrid work and reimagine everything from meetings that transcend space and time to a digital employee experience that everyone can access from anywhere – right in the flow of their work.   

YouTube Video

Helping  every organization have better  hybrid meetings  

Right now, every leader is grappling with how to have effective hybrid meetings, often with limited equipment and resources. Hybrid meetings are an entirely new kind of meeting and require rethinking how we approach and conduct them to ensure we’re putting every attendee on equal footing, no matter where they’re sitting.   

We can all start making hybrid meetings better today with a few good habits, like making sure in-person meeting rooms have a centralized audio device; encouraging everyone to join the Teams meeting – with the camera on – and appointing an in-room moderator to facilitate conversation by monitoring the chat, raise hands feature, and participants unmuting themselves to speak.  

New cultural meeting norms are key, but technology can help take the hybrid meeting experience to the next level. Today, we’re announcing new Microsoft Teams Rooms innovations, along with new Teams features, to help you have impactful, engaging hybrid meetings where everyone feels included.   

New Teams Rooms innovations  

The next wave in Teams Rooms innovation will include a new category of AI-enabled cameras. There are three unique technologies that power this new category of intelligent cameras. The first is  AI-powered active speaker tracking , enabling in-room cameras to use audio, facial movements and gestures to detect who in the room is speaking – zooming in for a closer perspective. The next is  multiple video streams  that allow in-room participants to be placed in their own video pane. The third is  people recognition , which will identify and display the profile name of enrolled users within their video pane.   

Screenshot of Teams video meeting

New hybrid meeting features  

To make remote presentations more engaging, we’re announcing  cameo,  a new PowerPoint experience that integrates the Teams camera feed into a presentation to allow the presenter to customize how and where they want to appear on the screen with their slides. As we work to create an inclusive hybrid work culture, where all voices can be heard,  speaker coach  in Microsoft Teams uses AI to privately share guidance on your pace, notify you if you are interrupting someone and remind you to check in with your audience.  

Planning and understanding who will be attending and how they will attend is also key ,  and today we are announcing a new  Outlook RSVP  feature that lets people specify whether they will attend in person or remotely. The redesigned  Working hours  will allow you to include work schedule specifics directly in your calendar, so others can know when and where you’ll be working as we adjust to more flexible work hours and locations.   

Cameo screenshot

A digital employee experience – right in the flow of work  

Going forward, the digital employee experience  is  the employee experience. Today, we are announcing the public preview of the Microsoft Viva Connections mobile app, available later this month. Viva Connections provides a single employee app for company communications, news and announcements – all in a personalized feed right in Teams, powered by Microsoft 365. Organizations can customize the dashboard with additional apps to empower employees, making it easier to find useful resources and complete tasks like submitting expense reports, requesting time off from work, or clocking in and out. Powered by adaptive cards, dashboard items can be targeted to specific groups of employees so everyone can stay connected and engaged wherever and however they work.  

Our already available  Return to the Workplace solution, built on the Power Platform, gives organizations a set of Power Apps and Power BI dashboards to keep employees safe – including vaccine attestation, case management, health screening, facility safety management and more. It’s a turnkey solution that can also seamlessly integrate with existing third-party apps an organization may already have in place. All these solutions are easily incorporated into the Viva Connections experience.   

Flexible, safe workspaces for a hybrid workforce  

With an increasingly mobile workforce, it’s important for organizations to be able to offer a space for employees who need workspaces where they can touch down either for the day or just a couple of hours to get work done. We’re announcing a  new hotdesking  experience on Microsoft Teams displays that allow people to locate and reserve flexible workspaces in the office. Employees can book a space from the device or in advance using Outlook or Teams and access their personal Teams calendar, chats, meetings and more. Teams displays can be used as a stand-alone device or as a second screen when hot desking. When an employee signs out, all personal information will be removed from the device.  

For more of today’s Teams announcements, please check out  the Microsoft 365 blog .    

Giving people more choices in how they want to work  

To help employers and hiring managers adapt to this new world of work, LinkedIn is rolling out new fields within job postings where organizations can now signal if the open job is remote, hybrid or on-site, helping job seekers search and discover jobs that align with how they want to work. And coming soon,  LinkedIn will also have a way for companies to share how they are approaching the future of work on their company page including vaccination requirements, if they plan to go back to an office, stay remote or go hybrid.  

Free LinkedIn Learning courses   

Investing in new skills is crucial to thriving in today’s workplace –  be it in-person, remote or hybrid. To help, LinkedIn has made nearly 40 LinkedIn Learning courses free until Oct. 9, 2021 so that LinkedIn members and HR pros can build critical skills. Whether you’re looking to  support your team through the transition to hybrid work  or be the  kind of manager no one wants to leave , these courses can help you lead through change.   

An intelligent skill-building platform to help employers invest in their greatest asset  – their people  

People are rediscovering how they find meaning and purpose in their lives and at work, and responding by seeking out new opportunities. Leaders, meanwhile, are facing new challenges in retaining and reskilling their people to meet the demands of our new world of work. To navigate these changes, LinkedIn’s skill-building platform, LinkedIn Learning Hub, will be generally available in two weeks. It will help companies build their employees’ skills through personalized content, community-based learning and skill-development insights that will  i nform their overall learning and development strategy.

Screenshot of LinkedIn Learning Hub

Our commitment to customers around the world  

Meeting employee expectations, not just to attract and to retain talent, but to enable personal well-being, will be a challenge for every leader and every organization. But if we embrace the data, listen to our employees and customers and incorporate flexibility into everything we do, we believe we can create a better future of work. At Microsoft and LinkedIn, we’ll continue to use what we’re learning to build and improve products designed to empower people for the ways we work today and the ways we’ll work in the future. And we’ll continue sharing what we learn along the way.   

Tags: hybrid workplace , LinkedIn , Microsoft 365 , Microsoft Teams , Microsoft Viva

  • Check us out on RSS

microsoft research articles

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

Preview improvements coming to the PMC website in October 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Front Big Data

Logo of frontbigdata

A Review of Microsoft Academic Services for Science of Science Studies

Associated data.

All datasets generated for this study are included in the article/ Supplementary Material .

Since the relaunch of Microsoft Academic Services (MAS) 4 years ago, scholarly communications have undergone dramatic changes: more ideas are being exchanged online, more authors are sharing their data, and more software tools used to make discoveries and reproduce the results are being distributed openly. The sheer amount of information available is overwhelming for individual humans to keep up and digest. In the meantime, artificial intelligence (AI) technologies have made great strides and the cost of computing has plummeted to the extent that it has become practical to employ intelligent agents to comprehensively collect and analyze scholarly communications. MAS is one such effort and this paper describes its recent progresses since the last disclosure. As there are plenty of independent studies affirming the effectiveness of MAS, this paper focuses on the use of three key AI technologies that underlies its prowess in capturing scholarly communications with adequate quality and broad coverage: (1) natural language understanding in extracting factoids from individual articles at the web scale, (2) knowledge assisted inference and reasoning in assembling the factoids into a knowledge graph, and (3) a reinforcement learning approach to assessing scholarly importance for entities participating in scholarly communications, called the saliency, that serves both as an analytic and a predictive metric in MAS. These elements enhance the capabilities of MAS in supporting the studies of science of science based on the GOTO principle, i.e., good and open data with transparent and objective methodologies. The current direction of development and how to access the regularly updated data and tools from MAS, including the knowledge graph, a REST API and a website, are also described.

Introduction

Centuries of scientific advancements have been a result of a virtuous cycle where scientists meticulously collect observation data to deduce a theoretical model and then use the model to predict new experimental outcomes as a means to validate the theory. This scientific principle has been applied to study the science of science, namely, the development of science itself, a field that sees pioneers like Eugene Garfield at the Institute for Scientific Information (ISI, now part of Clarivate Analytics) (Garfield, 1955 , 1964 , 1972 ). Driven by the insights that scientific advancements inevitably leave abundant traces in the scholarly communications that often manifest themselves in the form of citations, a central topic in the science of science has been deriving quantitative models from citations for the purpose of analyzing and understanding the impacts of scientific work. Historically, citations made in the main body of an article have been difficult to collect so the bibliography has been used in their stead. Implicitly, this practice assumes the relations among publications can be approximated by the pairwise Boolean measures between the citing and the cited articles. Such an approximation is found to be too reductive in contrast to peer reviews for article-level assessments (Wilsdon, 2015 ), although there is evidence suggesting noises in such a simplified model may be “canceled out” through aggregations at a level higher than individual articles (Traag and Waltman, 2019 ). Indeed, the most widely used bibliometrics, such as the journal impact factor (JIF) or the h-index, are by design aggregate measures at the journal or the author level. However, the demands for article-level metrics are so strong that they make popular a practice assuming articles in the same journal are equal in quality and the aggregate metrics for the journal can serve as a proxy for the articles published therein. Its adverse effects are so profound and misuses so pervasive that renowned institutions and thought leaders have found it necessary to proclaim the San Francisco Declaration of Research Assessment (DORA) 1 to publicize a strong stance against using journal-level metrics for research assessments. A widely accepted good model to understand the impacts of individual publications has yet to be found.

Another challenge in the study of science of science is the explosive growth in the volume of scientific reports and the diversity of research topics. These have outstripped the cognitive capacity of human beings to properly digest and catch up. This cognitive overload ostensibly impacts everyone, including those employed by vendors to curate data and develop commercial platforms for science of science studies. As a result, errors and omissions in manually curated data are abundant, eroding the trustworthiness of studies conducted on those platforms. Most frustratingly, the proprietary and opaque nature in the commercial systems prevent recourses when obvious errors are spotted. As data-driven decision-making processes have become more prevalent in recent years, the platform quality has become a serious issue that prompts the Computing Research Association (CRA) to release a stern statement on the worsening state of commercial data and call for actions against unscientific practices based on or leading to flawed data 2 . In their report (Berger et al., 2019 ), a CRA working group illustrates faulty data from Clarivate Analytics and surveys from humans no longer up to date in their fields may have led US News & World Report to produce abhorrent rankings on research universities that can severely mislead students in making school choices and funders in allocating resources. Similar to DORA, the CRA working group publishes a set of guidelines urging the adoption of Good and Open data with Transparent and Objective methodology, known as the GOTO principle, in conducting and publishing the results of quantitative science of science studies.

This article describes Microsoft Academic Services (MAS), a project in Microsoft Research with an aim to support researchers to follow the GOTO principle. Having evolved from the initially disclosed in (Sinha et al., 2015 ), MAS now consists of three parts: an open dataset known as Microsoft Academic Graph (MAG) 3 , a freely available inference engine called Microsoft Academic Knowledge Exploration Service (MAKES), and a website called Microsoft Academic 4 that provides a more human friendly interface to MAKES. MAS is a part of an ongoing research that explores the nature of cognition, a topic in artificial intelligence (AI) that studies the mental capacity in acquiring, reasoning and inferencing with knowledge. The research is motivated by the observation that cognition involves the capabilities of memorizing, computing, being attentive, and staying focused on the task at hand, all of which can be programmed to the modern computer to outperform humans. Particularly for MAS, the project explores the boundary within which the power of machines can be harnessed to understand the scholarly communications observable on the web. In other words, MAS aims at developing AI agents that are well-read in all scientific fields and hopefully can become trustable consultants to human researchers on matters of scholarly activities taking place on the web. In this sense, the MAG component in MAS is the outcome of the knowledge acquisition and reasoning and MAKES, the capability of machine inferencing with the knowledge in MAG. The dataset MAG is distributed and frequently updated under an open data license and the inference algorithms in MAKES are published in relevant peer-review venues and summarized later in this article.

Aside from being open in data and transparent in algorithm as per the GOTO principle, MAS actively uses technologies to capture scholarly communication activities with adequate quality and coverage to strive for a good platform. To address the explosive growth in scientific research, MAS employs the state-of-the-art AI technologies, such as natural language understanding, to extract the knowledge from the text of these publications. This allows MAS to always take a data-driven approach in providing consistent data quality and avoid manual efforts that are often the source of subjective controversies or errors. Knowledge extraction in MAS goes beyond simply indexing key phrases to recognize and disambiguate the entities underpinning scholarly communications. MAS currently includes entities that describe who supported by which institutions have made what claims in which publication at which instance of which venue , as illustrated in Figure 1 . With more scholarly communications being conducted online with data and software tools, the definition of publication in MAS has been expanded. Aside from the traditional forms such as books, journals and conference papers, MAS has recognized datasets and software packages as additional forms of publications. Additionally, as plenty of scholarly work exerts impacts through commercial exploitation preceded by patent applications, MAS has also included them as publications. These new resources fit well into the model of publication entity in Figure 1 because they all have authors, affiliations, topical contents, etc., and can receive citations. In addition to extracting these entities, a key mission of knowledge extraction is to recognize the relations among the entities, such as the citation contexts characterizing how the work in one publication is received by others citing it. As schematized in Figure 1 , these entities and their relations are represented in a graph structure as the nodes and edges, respectively, leading to the name of MAG. Note that the entity recognition and disambiguation (ERD), as reported in (Carmel et al., 2014 ), is far from a solved problem. However, the key here is the AI technologies employed in MAS are designed to learn and improve by itself by repeatedly reading more materials than any human can possibly do in a lifetime. After years of self-improving, many independent studies have suggested that MAG data are in many aspects as accurate, if not more, than manually curated data (Herrmannova and Knoth, 2016 ; Harzing and Alakangas, 2017 ; Hug and Brändle, 2017 ; Hug et al., 2017 ; Thelwall, 2017 , 2018a , b , c ; Kousha et al., 2018 ).

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0001.jpg

The data model of scholarly communications in MAS where the nodes represent the entity types modeled in MAG, and the simple and block arrows depict one-to-one and one-to-many relations among the entities, respectively.

Secondly, MAS uses technologies for scale, particularly when the lack of coverage in many datasets is becoming ever more concerning. While it might be appropriate in the last century for human experts to manually select only some of the scholarly communications into a database, this practice may have finally lived out its usefulness as the case studies in the CRA report have shown. Furthermore, with the advancements in information technology, online publishing has become a widely adopted medium for scientists to communicate with one another. Important activities, including self-archiving, data and software sharing, and community efforts dedicated to reproducing previously published results [e.g., Papers with Code 5 , ReScience (Rougier et al., 2017 )] are taking place exclusively on the web. A modern dataset therefore must be able to capture all these web-only activities to properly reflect the current state of the reality, and it is hard to fathom how all these capturing efforts can be accomplished by hand. MAS provides an encouraging example that technologies can help in this area.

The key to MAS is large-scale deployment of AI agents in understanding scholarly communications. Therefore, the rest of the article is devoted to describing the methodologies so that the characteristics of MAS can be better understood. The AI technologies used in MAS, as illustrated in Figure 2 , encompass three areas: (1) natural language understanding, including ERD and concept detection to extract factoids from individual publication and to fulfill queries in MAKES, (2) knowledge reasoning to organize the factoids into MAG, and (3) a reinforcement learning system to learn a probabilistic measure called the saliency that facilitates the statistical learning and inferences in the above two areas.

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0002.jpg

AI and service components in MAS are comprised of two feedback loops, one to grow the power of acquiring knowledge in MAG and the other to assess the saliency of each entity in MAG. In the first loop, each publication on the web is first processed by the MAG assisted entity recognition and disambiguation as described in (1). As the raw entities and their relations are extracted from individual publications, semantic reasoning algorithms are then applied to conflate them into a revised graph, including the concept hierarchy from all the publications. The revised MAG is then used in the next run to better extract entities from publication. The second loop utilizes the citation behaviors as the rewarding target for a reinforcement learning algorithm to assess the importance of each entity on MAG based on the network topology. The quantitative measure, called the saliency, serves as a ranking factor in MAKES, a search and recommendation engine for MAG.

Entity Recognition and Disambiguation

Central to MAS is the quest to harness the power of machine to acquire knowledge from written text. As alluded previously, the knowledge acquisition task amounts to recognizing the lexical constructs of the semantic objects representing either entities or relations. To be more precise, the task of natural language understanding in MAS is formulated as a maximum a posteriori (MAP) decision problem:

where the input x = ( w 1 , w 2 , ⋯ ) is a word sequence of a natural language expression, K is a knowledge base, and the task is to find the best output ŷ = ( e 1 , e 2 , ⋯), e i ∈ K , that is a sequence of semantic objects. For example, suppose the input is a sentence “HIV causes AIDS.” The ideal output should consist of two entities “HIV” and “AIDS,” and a relation “causing” between them.

The MAP decision is known to be optimal provided the posterior probability distribution in (1) can be accurately estimated. While this can be done directly, MAS uses a mathematically equivalent approach, known as the generative modeling, where the Bayes rule is applied to (1) to rewrite the MAP decision as:

with P ( x | y, K ) and P ( y | K ) the semantic language and the prior models, respectively. The semantic language model characterizes how frequently a sequence of semantic objects y is expressed through the word sequence x . Typically, an entity is lexicalized by a noun phrase while a relation, a verb phrase. MAS, however, does not utilize the syntax structure of natural language but, rather, just assumes that the lexical realization of each semantic object is statistically independent of one another, namely:

where x i denotes the i -th phrase segment in x corresponding to e i . Essentially, the semantic language model characterizes the synonymous expressions for each semantic object e i and how likely each of them is used. For example, the journal “Physical Review Letters” can be referred to by its full name, a common abbreviation “Phys Rev Lett,” or simply the acronym “PRL,” and an author can be mentioned using the last name, the first name or just its initial with an optional middle initial. The bibliography section, the text body and the web pages of a paper all provide abundant materials to harvest synonymous expressions. With large enough data samples, it appears adequate in MAS to use a simple maximum likelihood estimation, i.e., frequency counts with statistical smoothing, for the synonym model P (· | e i , K ).

The semantic prior model P ( y | K ) assesses the likelihood of a certain combination of semantic objects that can be derived from the knowledge base. In a way, the brunt of the statistical independent assumption in (3) is lessened because the contextual dependencies leading to a viable semantic interpretation are strictly enforced here. This can be seen by applying the chain rule of conditional probability to further decompose the semantic prior model as:

where P ( e 1 | K ) is the saliency of the entity e 1 and P ( e i | e i −1 , ⋯ e 1 , K ) is the semantic cohesion model according to the knowledge K . In conjunction with the synonym model, the semantic cohesion model can be estimated directly from data with an additional constraint that assigns zero probability to implausible semantic object combinations. This constraint plays a critical role in reducing the degree of ambiguities in understanding the input. For example, “Michael Evans” with a missing middle initial is a very confusable name, and “WWW” can mean a conference organized by IW3C2, a journal (ISSN: 1386-145X or 1573-1413), or even as a key word in the title of a paper. However, there are only two authors, a “Michael P. Evans” and a “Michael S. Evans” that have ever published any papers in the WWW conference, in 2002 and the other in 2017, respectively, and never in the namesake journal or any paper containing “WWW” as a key term in all other publication venues. If the publication year is also present, the apparently ambiguous input “Michael Evans (Evans and Furnell, 2002 )” can be precisely resolved into the entity referring to the author named “Michael P. Evans” that has published a paper in “the eleventh International World Wide Web Conference” held in Honolulu Hawaii in the year of 2002. Using the knowledge-imposed constraints is particularly effective for author disambiguation when the technique is applied to understand curricula vitae or author homepages posted on the web. Assuming each such web page belongs to a single author, the publications listed therein are often high-quality signals to ascertain the identity of the author from the namesakes.

The manner that the knowledge is utilized in (4) also allows MAS to identify and acquire new synonymous expressions for existing entities and, often, new entities. This capability in acquiring new knowledge without human intervention is the key for MAS to enrich itself gradually. Mathematically, let K t denote the knowledge base used in (1) leading to the understanding of the scholarly materials y t at time t , the knowledge enrichment in MAS at an interval of Δ t , is also formulated as the MAP decision:

The iterative process of (5) in MAS can be better appreciated through a common task in parsing the bibliography where the author intent is to refer to a publication with a sequence of references to authors, followed by an optional publication title and a reference to a publication venue. The manners with which a reference is made, however, is highly inconsistent. When the semantic knowledge is applied to parse an input such as “Zhihong Shen, Hao Ma, and Kuansan Wang, ACL-2018,” it allows MAS to recognize fragments of the text, say, “Hao Ma” or “Kuansan Wang” as authors as they are frequently seen in the knowledge base. With these anchors, MAS can use (4) to infer that “Zhihong Shen” and “ACL-2018” are likely references to another author and the venue, respectively. These inferences can be made even before the publication records of ACL-2018 are included into the knowledge base and can be used with (5) to grow new entities in MAS.

While MAG only publishes the canonical expression for each entity, MAKES includes the probabilistic models that we derive from all the raw materials mentioned above. A step-by-step examination of (2) can be conducted in the query input box at the Microsoft Academic website where, upon each character entered, an API call is made into MAKES to analyze the semantic intent of the typed input with the MAP decision rule described in (2). Top interpretations manifesting themselves as query completions or suggestions are displayed to the user as a means for query intent, disambiguation or confirmation. More details are described in the FAQ page of the website 6 .

Concept Detection and Taxonomy Learning

Like many complex systems, the relations among entities in the scholarly communications ( Figure 1 ) cannot fully capture the activities because the semantics of the communications is encoded not in the topology but in the natural language contents of the publications. To address this issue, MAS adopts an entity type, called concepts [called “fields of study” in Sinha et al. ( 2015 )], to represent the semantic contents of a document. Unlike physical entities such as authors and affiliations, concepts are abstract and hence have no concrete way to define them. Furthermore, concepts are hierarchical in nature. For example, “machine learning” is a concept frequently associated with “artificial intelligence” that, in turn, is a branch of “computer science” but often intersects with “cognitive science” in “psychology.” Accordingly, a taxonomy must allow a concept to have multiple parents and organize all concepts into a directed acyclic graph (DAG). While concepts can be associated with all types of physical entities, say, to describe the topics of interest of a journal or the fields of expertise of a scholar, MAS only infers the relations between a publication and its concepts directly and leaves all others to be indirectly aggregated through publications.

A survey on the concepts taxonomy used in major library systems, presumably developed by human experts, suggests that few of them are compatible with each other. The low agreement among human experts leads MAS to create a concept taxonomy by itself solely from the document collection. As there are close to 1 million new publications a month being added in recent months, the machine learned taxonomy is dynamically adjusted on a regular basis so that new concepts can be added and obsolete concepts can be retired or merged with others.

Concept detection is a natural language understanding problem and, therefore, its mathematical foundation is also governed by (1). Unlike the ERD problem, however, the ideal output y ^ in this case is an unordered collection of DAGs of concepts rather than a sequence of semantic objects, and the textual boundaries of a concept in x are intrinsically soft, i.e., phrase segments can overlap. MAS therefore employs the approach to directly estimate the probabilistic distribution in (1) from the text rather than going through a generative model of (2). As detailed in the recent publication (Shen et al., 2018 ), the key concept underlying the MAS approach here is the distributional similarity hypothesis proposed in 1950's (Harris, 1954 ), which observes that semantically similar phrases tend to occur in similar contexts. There have been plenty of methods reported in the literature demonstrating the efficacy of applying distributional similarity for concept detection, either by training a hierarchical classifier mapping a sequence of discrete words directly into concepts, or by the embedding method that first converts the text into a vector representation with which learning and inferences can be conducted in a vector space (Turney and Pantel, 2010 ). When properly executed, semantically similar phrases can be transformed into vectors close to one another, simplifying the synonymous expression detection, needed for (3), into a nearest neighbor search. In other words, the probabilistic distribution of synonyms P (·| e i , K ) can be estimated by the distance in the vector space. Recently, the embedding methods have produced many surprising results, starting with (Mikolov et al., 2013 ; Berger et al., 2019 ), that contribute to a renaissance of the vector space model thanks to the availability of big data and powerful computational resources. The current practice in MAS, however, has found it more powerful to combine both the discrete and the vector space approaches into a mixture model for concept learning (Shen et al., 2018 ).

The concept detection software in MAS has been released as part of the MAG distribution. The package, called Language Similarity 7 , provides a function with which the semantic similarity of two text paragraphs can be quantified using the embedding models trained from the publications in the corresponding MAG version. This function in turn serves as a mixture component for another function that, for any paragraph, returns a collection of top concepts detected in the paragraph that exceed a given threshold. Again, interested readers are referred to the recent article (Shen et al., 2018 ) for technical details.

Network Semantics Reasoning

As MAS sources its materials from the web notorious for its uneven data qualities, duplicate, erroneous and missing information abounds. Critical to MAS is therefore a process, called conflation, that can reason over partial and noisy information to assemble the semantic objects extracted from individual documents into a cohesive knowledge graph. A key capability in conflation is to recognize and merge the same factoids while adjudicating any inconsistencies from multiple sources. Conflation therefore requires reasoning over the semantics of network topology and many of the techniques in MAS described in Sinha et al. ( 2015 ) are still in practice today.

Recently, a budding research area focuses on extending the notion of distributional similarity from its natural language root to the network environment. The postulation is straightforward: similar nodes tend to have similar types of edges connecting to similar nodes. Similar to the natural language use case of representing entities and relations as vectors, the goal of this approach is to transform the nodes and edges of a network into vectors so that reasonings with a network can be simplified and carried out in the vector space with algebraic mathematics. Network semantics, however, is more complicated than natural language whose contextual relations are single dimensional in nature: a phrase is either left or right to another. A network has a higher order topology because a node can simultaneously connect to a wide variety of others with edges representing distinctive relations. Citation network is a simple example where one paper can be cited by two others that also have a citation relation between them. Citation network is considered simple as it only has a single type of nodes, publication, and a single type of relation, citing. In reality, scholarly communications also involve people, organizations, locations, etc. that are best described by a heterogeneous network where multiple types of nodes are connected by multiple types of edges, making the notion of distributional similarity more sophisticated. The research in heterogenous network semantics reasoning, especially in its subfields of network and knowledge graph embedding, is ongoing and highly active.

MAS has been testing the network embedding techniques on related entity recommendation and found it essential for each entity to have multiple embeddings based on the types of relations involved in the inferences. In other words, embedding is sensitive to the sense defining similarity. For example, two institutions can be regarded as similar because their publications share a lot in common either in contents, in authorships, in venues, or are being cited together by same publications or authors. The multitude of senses of similarity leads to multiple sets of embeddings, of which results are included in MAG distributions. As the research in this area is still ongoing and the techniques by no means matured, MAS applications can achieve better results by combining the embedding and the discrete inference techniques. One such example is reported in a recent paper (Kanakia et al., 2019 ) that describes the method behind the current related publication recommendation in MAS. The user studies in this application show the best system uses both the distance of the text embeddings and the frequency of being cited together.

Assessing Entity Importance With Saliency

As the MAP decision in (1) also drives MAKES to rank the results y in response to a query x , the entity prior P ( e | K ) in (4) is a critical component for MAKES. The way the entity prior is estimated determines in which sense the ranking is optimized. Ideally, the prior should be the importance the entity has been perceived by the scholarly community in general. Recently, a new area of research, lumped under the name altmetrics (Piwowar, 2013 ), has been advocating that the searching, viewing, downloading, or endorsement activities in the social media for a publication should be included in estimating the importance of the scholarly work. Having monitored these activities for the past few years, we have found altmetrics a good indicator gauging how a publication has gained awareness in the social media. Although being known is a necessary step for being perceived as important, our observations cannot exclude the possibility that a publication is searched and viewed more because it is repeatedly mentioned in another highly regarded work, or authored by influential scholars or even just from reputable organizations. Based on our observations and concerns about altmetrics in the community (e.g., Cheung, 2013 ), the current focus in MAS is on exploiting the heterogeneity of scholarly communications mentioned above to estimate the entity prior by first computing the importance of a node relative to others of the same type and then weighting it by the importance of its entity type.

Saliency: An Eigencentrality Measure for Heterogeneous Dynamic Network

The eigenvector centrality measure, or simply eigencentrality, has been long known as a powerful method to assess the relative importance of nodes in a network (Franceschet, 2011 ). Developed in the early twentieth century, eigencentrality measures the importance of a node relative to others by examining how strongly this node is referred to by other important nodes. Often normalized as a probabilistic measure, eigencentrality can be understood as a likelihood of a node being named as most important in a survey conducted on all members in the network. The method is made prominent by Google in its successful adaptation of eigencentrality for its PageRank algorithm: the PageRank of a webpage is measured by the proportional frequency of the incoming hyperlinks weighted by the PageRank of the respective sources. In a distinct contrast to simple citation counts, two important considerations in PageRank are the frequency of mentions in the citing article counts, and the importance of the citing source matters. Google has demonstrated that PageRank can be successfully used to assess the importance of each web document.

There are, however, two major challenges in using the eigencentrality as an article-level metric in general. First, the eigencentrality is mathematically well-defined only if the underlying network is well connected. This mathematical requirement is often not met in real-life, neither in the citation networks nor the web graph. To tackle this problem, Google introduced a “teleportation” mechanism in PageRank in which the connection between two web pages is only 85% dependent on the hyperlinks between them. The rest of the 15%, called the teleportation probability, is reserved for the assumption that all webpages are connected to each other intrinsically and uniformly. While the teleportation mechanism serves Google well, it is found to be fragile and implausible for the citation network (Walker et al., 2007 ; Maslov and Redner, 2008 ): the ranking of scholarly publications is overly sensitive to the choice of the teleportation probability, and the best choice suggests scientists only follow the bibliography half the time, with the other half randomly discovering articles from the entire research literature following a uniform distribution. Many PageRank inspired studies, as recently reviewed in (Waltman and Yan, 2014 ), have also made the same observation and proposed remedies utilizing the heterogeneity of the scholarly communication network. They mostly, however, are in an early exploratory stage as the manners in modeling the heterogeneous interactions still contain many heuristics needed to be further validated. Secondly, even the well-connected issue can be addressed through a heterogeneous model, another challenge, as pointed out by many (e.g., Walker et al., 2007 ), is how to avoid treating eigencentrality as a static measure so that the time differences in citations can be taken into account. It is undesirable to treat an article that receives the last citations long ago as equal to one that has just received the same amount of citations today because results without a proper temporal adjustment exhibit a favorable bias toward older publications that have more time to collect citations.

MAS attacks these two challenges with a unified framework called saliency based on the following considerations. First, to address the underlying network as changing in time, saliency is defined as the stochastic process characterizing the temporal evolution of the individual eigencentrality computed from a snapshot of the network. Without making assumptions on its form, the autoregressive moving-average (ARMA) process, mathematically known to be able to approximate a non-stationary distribution to any precision with enough orders, is used to model the temporal characteristics of saliency. Surprisingly for MAS, a simple first order autoregressive (AR) process seems sufficient for the model to reach an ergodic solution (to be shown below), suggesting that the endorsement power of a citation can be treated as simply as an exponential decay with a constant half-life interval. This finding is a validation of the observation first reported in (Walker et al., 2007 ).

Secondly, to account for the heterogeneity of the network, MAS uses a mixture model in which the saliency of a publication is a weighted sum of the saliencies of the entities related to the publication. By considering the heterogeneity of scholarly communications, MAS allows one publication to be connected to another through shared authors, affiliations, publication venues and even concepts, effectively ensuring the well-connectedness requirement is met without introducing a random teleportation mechanism. Mathematically, let s x ( t ) denote the saliency vector of the entities of type x at time t , with x = p specifically for the publication, the heterogeneous mixture model coupled with an AR process leads to:

where Δ t is the interval between the two successive network snapshots are taken, w p, x the (non-negative) weight of a type x node on the publication, τ the time decaying factor in the AR process, and A p, x the adjacency matrix characterizing the connection strength between a publication to any entity of type x . Currently, MAS considers all nodes of types x ≠ p to have equal connection to the publication, e.g., given a publication all of its authors and affiliations are treated as contributed equally to the saliency of the publication. In the meantime for publications citing one another, A p, p is set proportional to the number of mentions in the text body of the citing article to the cited work.

As the heterogeneous model treats the saliency of a publication as the combined saliencies of all entities related to it, s p ( t ) is therefore a joint probabilistic distribution. Accordingly, the saliency of a non-publication entity can be obtained by marginalizing the joint distribution, i.e.,

where A x, p = [δ ij ] and

Again, the current MAS implementation does not address how the credit of a publication should be assigned unevenly to its authors based on the author order as (8) implies all authors have equal contributions, the side effect of which, however, is each institutions associated with a publication will receive its credit proportional to the number of authors affiliated with the institution. Ostensibly, a more sophisticated model than (8) can be used where, for instance, the author sequence can play a role in determining δ ij . MAS reports the author sequence as well as the affiliation sequence for authors with multiple affiliations, but has not yet used them for the purpose of computing saliency.

Estimating Saliency With Reinforcement Learning

To avoid making a strong assumption that the latent variables τ and w p, x are constant, MAS uses reinforcement learning (RL) to dynamically choose the best values based on the reinforcement signals streaming in through the observations. The choice is motivated by the fact that the RL technique is known to be effective in tackling the exploitation vs. exploration tradeoff, which in MAS means a balanced treatment between the older and newer publications or authors that have unequal time to collect their due recognitions. Often, the challenge of applying RL is the reinforcement signals are hard to obtain. This is fortunately not the case in MAS because approximately half a million new publications with tens of million citations are discovered every 2 weeks (= Δ t ), and these new observations provide ample materials to be reinforcement signals. Assuming the scholar communications are eventually just, namely, more important publications will receive higher citations in the long run ( NΔt, N ≫ 1), the goal of the RL in MAS is to maximize the agreement between the saliencies of today and the citations accumulated NΔt into the future. Currently, MAS uses the maximum mutual information (MMI) as the quantitative measurement for the agreement, namely, if c ( t ) denotes the vector of citation mention counts for all publication, the objective of the RL in MAS is to find:

where < ·, · > denotes the inner product. The choice of MMI allows (9) to be a convex function so that it can be iteratively solved with a quasi-Newton method. An off-the-shelve software implementing a L-BFGS algorithm is used in MAS.

It is a surprise that, by choosing long enough future N , the solutions to the latent variables τ and w p, x appear to be quite steady over time with the simplest form of ARMA process: 1 st order autoregression and no moving average. This apparent ergodicity allows MAS to administer the RL with a delay of NΔt ≈5 years, namely, the latent variables in (6) can be obtained by using the data observed up to 5 years ago to predict the citations of the recent 5 years. The results, as shown in Figure 3 , suggests that generally a publication accrues its saliency from citations at a weight slightly more than 92%, although the factors of its authors, affiliations, publication venues and even topics are non-trivial. Along the time domain, the value of τ, hovering around 0.9, corresponds to a temporal decay in saliency with a half-life of 7.5 years. In contrast to previously studies where the citations only account for 50% of the weight (Walker et al., 2007 ; Maslov and Redner, 2008 ) or with a very short decay from 1 to 2.6 years (Walker et al., 2007 ), the RL results in MAS are a much less dramatic departure from the common practice of using citation counts as an article level metric where, effectively, the metric is computed with a 100% weight in citations that do not decay over time.

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0003.jpg

The longitudinal values of the latent variables underlying the saliency as obtained by the reinforcement learning (RL) algorithm. These latent variables correspond to the weighting the algorithm has to exert on each entity type in order to predict the future citation behaviors most optimally in the sense of Maximum Mutual Information, as described in (9). The model shows citations remain the dominant factor to have a high saliency. Despite a relatively simple configuration, the model exhibits remarkable stability over the 35 months shown in this figure other than the two instances, in April 2017 and July 2018, when MAG changed its treatment on affiliations dramatically.

As also shown in Figure 3 , the RL is not impervious to major changes in the underlying data, such as the treatments to author affiliations. In May 2017, a so-called “inferred affiliation” feature was introduced to MAG where authors with unknown affiliations were associated with the “most likely” institutions inferred from their recent publication records. An overly optimistic threshold led to many inaccurate projections, and the RL responded to the degradation in quality by lowering the affiliation weight and shifting it to citations. In July of the following year, the MAG data schema is altered to allow a single author to have multiple affiliations and all of which receive equal attribution from the publications by the author. Such a more faithful characterization of author affiliations leads to a boost of the affiliation weight from 1.5 to 3.5%, suggesting the RL mechanism finds the affiliation information more useful.

Properties of Saliency

The saliencies obtained with (6) and (7) are reported in MAG at each update interval for all entities in a quantized form of −1000ln s ( t ), and are used as the entity prior in the MAP decision (1) in MAKES that can be examined through the search and analytics results at Microsoft Academic websites. All these tools can be valuable for more and deeper investigations to fully understand the properties of saliency as a potential metric. For example, by design s p ( t ) further discriminates the following three citation behaviors not considered in the simple citation count: the number of mentions in the citing article, the age of the citations received, and the non-citation factors that can alleviate the disadvantages for newer publications. The combined effects of these three aspects on the article-level assessment can be further studied by inspecting the results from (1) with synthesized queries. Figure 4 shows a typical outcome of 20% disagreement in the ranking position differences between saliency and citation count based rankings using the query set ( Supplementary Material S1 ). A quick examination into the disagreements confirms that a publication can have a higher saliency, albeit lower citation counts, because it is cited by more prestigious or more recent work as designed. Where these disagreements are desired, however, is a question worth exploring.

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0004.jpg

Histogram of ranking position by citation counts (CC) and saliencies (top) and their differences (bottom). Although future citation counts are the target for best estimating the saliencies, they only agree on the publication rankings roughly 80 percent of the time, demonstrating the effects of non-citation factors ( Figure 3 ) in the design of saliency. In contrast to citation counts, saliencies are sensitive to the venues, the authors, the concepts, and the recencies of the citing sources.

The design to unshackle the reliance on the overly reductive citation counts may also lead the saliency to be less susceptible to manipulations, ranging from citation coercions (Wilhite and Fong, 2012 ) to malicious cheating (López-Cózar et al., 2014 ) targeting metrics like the h-index. By using the citation contexts in saliencies, these manipulations are, in theory, less effective and easier to detect, as demonstrated by PageRank for the link spam detection in the web graph (Gyöngyi and Garcia-Molina, 2005 ). The extent to which the gain of the eigenvector-based method can be transported from the web graph to the scholarly network, however, awaits further quantification.

Another research topic MAS can be useful is in the effectiveness of saliencies of non-publication entities that, as described in (7), are aggregated from publication saliencies. This design gives rise to at least two intriguing properties. First, an entity can achieve high saliency with lots of publications, not all of which are important. As a result, saliency appears to be measuring both productivity and impact simultaneously, just like h-index. Indeed, shows a comparison between the h-index and the saliency of Microsoft authors ( Supplementary Material S2 ). Overall, there is a trend line suggesting individuals with a higher h-index tend to also have a higher saliency, but notable disagreements between the two abound. The author with the highest h-index, 134, in this set has the most publications at 619 articles that in total receive 60,157 citations but ranks only at the 4th place by saliency. Conversely, the highest ranked author by saliency has published only 138 papers receiving 82,293 citations with a h-index at 76. Most notably, the second highest ranked author by saliency has an h-index only at 31. This is because the author has published only 39 papers, which limits the h-index, but they are all well received with a total citation count of 58,268, which buoys the saliency. The drawbacks of the h-index, e.g., capping at the publication counts in this example, are well-known (Waltman and Eck, 2012 ). By considering more factors and not limited to overly reductive raw signals, saliency appears to be better equipped to avoid mischaracterizing researchers who strive for the quality and not the quantity of their publications.

Secondly, because the underlying foundation of an aggregated saliency is based on article-level analysis, interdisciplinary work seems to be better captured. One such example is the journal ranking on a given subject, say, Library Science. As shown for a while at Microsoft Academic website 8 , journals like Nature and Science are among the top 10 for this field when ranked by saliency. This may be a surprise to many human experts because these two journals are seldom considered as a publication venue for the field of library science. Indeed, if the journals are ranked by h-index, these two journals will appear in much lower positions because the numbers of articles in the field are lower in these two journals. However, a closer investigation shows that these two journals have influential articles in the field, such as the Leiden Manifesto (Hicks et al., 2015 ) in Nature and the coercive citation studies (Wilhite and Fong, 2012 ) in Science. If one were to understand the most impactful papers in this field, precluding these two journals into consideration would lead to unacceptable omissions and result in incomplete work. Again, this example highlights the known problem of using journals as the unit to conduct quantitative scientific studies, and the sharp focus into article-level analysis, as demonstrated feasible by saliency, appears to be a better option.

Prestige: Size-Normalized Saliency

A known issue existing in aggregate measurements is that the sheer number of data points being considered can often play an outsized role. This can be seen in Figure 5 where the author saliency largely agrees, especially for prolific authors, with the h-index, a metric designed to measure the impact as well as the productivity. As implied by (7), an author can reach a high saliency by having a large number of publications despite most of them receive only moderate recognitions. Given it has been observed that hyper-prolific authors exist (Ioannidis et al., 2018 ), and their publications seem to yield uneven qualities (Bornmann and Tekles, 2019 ), it might be helpful to juxtapose the saliency with a corresponding size-normalized version, which we call prestige, to further discern the two aspects. To be specific, the prestige of a non-publication entity can be derived from (7) as:

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0005.jpg

A scatter plot comparing the h-index and the saliency where each dot corresponds to the h-index and the saliency of a Microsoft author. Although the two metrics largely agree, the saliency measure is able to overcome a known limitation of the h-index and highlight authors who have published widely recognized work but not in large quantity.

where A ¯ x , p = [ δ i j / ∑ j δ i j ] in contrast to (9). In short, the prestige of an entity is the average of the saliencies of its publications. Figure 6 illustrates the effect of the size normalization through the rankings of the world research institutions in the field of computer science based on the saliencies and the prestiges of their research papers published during the 5-year window between 2012 and 2016. The institutions that publish with consistent impacts are lined up along the main diagonal where the size normalization has negligible effect on their rankings. It appears the majority of the institutions are in this category. Scattered to the upper left of the diagonal, however, are those that are not the most prolific institutions but, when they do, their publications tend to be highly recognized by the research community. Size normalization, as expected, significantly boosts their rankings, for instance, as in the casesof Princeton University and Google. On the other hand, clustered to the lower right to the diagonal are the institutions that achieve high saliencies by publishing a large body of literature, as reflected in the relatively large bubble sizes in Figure 6 , and hence their rankings are negatively impacted by the size normalization operation.

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0006.jpg

Saliency (horizontal) vs. Prestige ranking of the top 50 research institutions in computer science area. The size of each bubble corresponds to the number of publications included in computing the saliency and the prestige measurements.

As many GOTO-compliant ranking systems (Berger et al., 2019 ) have discovered, one cannot over-emphasize that institution rankings are a highly sophisticated task that necessitates multiple perspectives and with varying degrees in granularities that commercial rankings such as US News & World Report are typically ill-equipped. To illustrate the point, Figure 7 shows the saliency-prestige rankings of institutions in the subfield of computer science, artificial intelligence, and its subfields of machine learning, computer vision and natural language processing. The recurring themes emerging from the high variances in the ranking results and significant differences in the top institutions strongly suggest that the ranking result of a field is a very poor predictor of its subfields. This is consistent with our observation that, within the subfields of computer science, the spectrum of research topics is so broad that institutions can choose to specialize into a selective few to have a strong and highly impactful research program. Consequently, ranking institutions at too broad a category amounts to comparing research on notably different fields that can have distinct publication culture and citation behaviors, i.e., is an apple-vs.-orange type of comparison. With new resources like MAS that can pinpoint each publication to very fine-grained fields of study, such a deeply-flawed methodology that was previously tolerated due to data scarcity should no longer be deemed acceptable and must be soundly rejected by the community.

An external file that holds a picture, illustration, etc.
Object name is fdata-02-00045-g0007.jpg

Institution Rankings, by Saliency (horizontal) vs. Prestige, for the field of Artificial Intelligence and its subfields.

The explosive growth in scholarly communications has made it more difficult for individual humans to keep track of the latest achievements and trends in scientific research. The warning signs are visible in the worsening qualities of research assessments involving expert opinions, as a recent CRA study showed. This article describes how MAS utilizes the advancements in AI to curate a good and open data set and enable transparent and objective methodologies (GOTO) for scientific studies on science. The AI components in MAS, in natural language understanding, in knowledge reasoning and inferences, and in reinforcement learning for estimating saliencies of entities in scholarly communications have been described. There are early indications that saliencies, an objective measure by harvesting the peer reviewed citation contexts, avoid many drawbacks of existing academic metrics.

Data Availability Statement

Author contributions.

KW drafted the manuscript and coordinated the research project. ZS, RR, and DE supervised the MAG, MAS, and the MAKES portions of the work. CH reviewed the experimental setups and software, while the rest of the authors have equal contributions to the data collected in the work.

Conflict of Interest

All authors are employed by the company Microsoft Research.

Acknowledgments

Dr. Hao Ma led the efforts in creating many advanced features in MAG. Dr. Bo-June Paul Hsu led the team to develop the inference engine in MAKES, and, with the assistance of Dr. Rong Xiao, implemented the first version of the reinforcement learning to compute saliency. The work would not be possible without the strong support from Microsoft Bing engineering teams and colleagues in Microsoft Research labs around the globe.

1 https://sfdora.org/

2 See https://cra.org/cra-statement-us-news-world-report-rankings-computer-science-universities/

3 https://docs.microsoft.com/en-us/academic-services/

4 https://academic.microsoft.com

5 https://paperswithcode.com/

6 https://academic.microsoft.com/faq

7 https://www.microsoft.com/en-us/research/project/academic/articles/understanding-documents-by-using-semantics/

8 See the analytic page at https://academic.microsoft.com/journals/41008148,161191863

Funding. The authors declare that this study received funding from Microsoft Research. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdata.2019.00045/full#supplementary-material

Supplementary Material S1

A randomly synthesized query sets to study the differences between the citation count based and the saliency based ranking behaviors.

Supplementary Material S2

Analytical script and the data to study the difference between the h-index and saliencies for authors.

  • Berger E., Blackburn S. M., Brodley C., Jagadish H. V., McKinley K. S., Nascimento M. A., et al.. (2019). GOTO rankings considered helpful . Commun. ACM 62 , 29–30. 10.1145/3332803 [ CrossRef ] [ Google Scholar ]
  • Bornmann L., Tekles A. (2019). Productivity does not equal usefulness . Scientometrics 118 , 705–707. 10.1007/s11192-018-2982-5 [ CrossRef ] [ Google Scholar ]
  • Carmel D., Chang M. W., Gabrilovich E., Hsu B. J. P., Wang K. (2014). ERD'14: entity recognition and disambiguation challenge, in Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval (Gold Coast, QLD: ). 10.1145/2600428.2600734 [ CrossRef ] [ Google Scholar ]
  • Cheung M. K. (2013). Altmetrics: too soon for use in assessment . Nature 494 , 176–176. 10.1038/494176d [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Evans M., Furnell S. (2002). A web-based resource migration protocol using WebDav, in Proceedings of the WWW-2002 (Honolulu, HI: ). [ Google Scholar ]
  • Franceschet M. (2011). PageRank: standing on the shoulders of giants . Commun. ACM 54 , 92–101. 10.1145/1953122.1953146 [ CrossRef ] [ Google Scholar ]
  • Garfield E. (1955). Citation indexes for science: a new dimension in documentation through association of ideas . Science 122 , 108–111. 10.1126/science.122.3159.108 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Garfield E. (1964). Science citation index- a new dimension in indexing . Science 144 , 649–654. 10.1126/science.144.3619.649 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Garfield E. (1972). Citation analysis as a tool in journal evaluation journals can be ranked by frequency and impact of citations for science policy studies . Science 178 , 471–479. 10.1126/science.178.4060.471 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Gyöngyi Z., Garcia-Molina H. (2005). Web Spam Taxonomy . Chiba: AIRWeb. [ Google Scholar ]
  • Harris Z. S. (1954). Distributional structure . WORD 10 , 146–162. 10.1080/00437956.1954.11659520 [ CrossRef ] [ Google Scholar ]
  • Harzing A. W., Alakangas S. (2017). Microsoft Academic: is the phoenix getting wings? Scientometrics 110 , 371–383. 10.1007/s11192-016-2185-x [ CrossRef ] [ Google Scholar ]
  • Herrmannova D., Knoth P. (2016). An Analysis of the Microsoft Academic Graph, D-lib Magazine 22 , 6. 10.1045/september2016-herrmannova [ CrossRef ] [ Google Scholar ]
  • Hicks D., Wouters P., Waltman L., Rijcke S. D., Rafols I. (2015). Bibliometrics: the leiden manifesto for research metrics . Nature 520 , 429–431. 10.1038/520429a [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Hug S. E., Brändle M. P. (2017). The coverage of Microsoft academic: analyzing the publication output of a university . Scientometrics 113 , 1551–1571. 10.1007/s11192-017-2535-3 [ CrossRef ] [ Google Scholar ]
  • Hug S. E., Ochsner M., Brändle M. P. (2017). Citation analysis with microsoft academic . Scientometrics 111 , 371–378. 10.1007/s11192-017-2247-8 [ CrossRef ] [ Google Scholar ]
  • Ioannidis J. P. A., Klavans R., Boyack K. W. (2018). Thousands of scientists publish a paper every five days . Nature 561 , 167–169. 10.1038/d41586-018-06185-8 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Kanakia S. Z., Eide D., Wang K. (2019). A scalable hybrid research paper recommender system for microsoft academic, in WWW '19 The World Wide Web Conference (New York, NY: ACM; ). 10.1145/3308558.3313700 [ CrossRef ] [ Google Scholar ]
  • Kousha K., Thelwall M., Abdoli M. (2018). Can Microsoft Academic assess the early citation impact of in-press articles? A multi-discipline exploratory analysis . J. Informet. 12 , 287–298. 10.1016/j.joi.2018.01.009 [ CrossRef ] [ Google Scholar ]
  • López-Cózar E. D., Robinson-García N., Torres-Salinas D. (2014). The Google scholar experiment: How to index false papers and manipulate bibliometric indicators . J. Assoc. Inform. Sci. Technol. 65 , 446–454. 10.1002/asi.23056 [ CrossRef ] [ Google Scholar ]
  • Maslov S., Redner S. (2008). Promise and pitfalls of extending Google's pagerank algorithm to citation networks . J. Neurosci. 28 , 11103–11105. 10.1523/JNEUROSCI.0002-08.2008 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Mikolov T., Sutskever I., Chen K., Corrado G. S., Dean J. (2013). Distributed representations of words and phrases and their compositionality, in NIPS'13 Proceedings of the 26th International Conference on Neural Information Processing Systems , Vol. 2 (Lake Tahoe, NV: ), 3111–3119. [ Google Scholar ]
  • Piwowar H. (2013). Altmetrics: Value all research products . Nature 493 , 159–159. 10.1038/493159a [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Rougier N. P., Hinsen K., Alexandre F., Arildsen T., Barba L. A., Benureau F. C. Y., et al.. (2017). Sustainable computational science: the ReScience initiative . PeerJ 3 , 1–8. 10.7717/peerj-cs.142 [ PMC free article ] [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Shen Z., Ma H., Wang K. (2018). A web-scale system for scientific knowledge exploration, in Meeting of the Association for Computational linguistics (Melbourne, VIC: ), 87–92. 10.18653/v1/P18-4015 [ CrossRef ] [ Google Scholar ]
  • Sinha A., Shen Z., Song Y., Ma H., Eide D., Hsu B. J. P., Wang K. (2015). An overview of Microsoft Academic Service (MAS) and applications, in Proceedings of the 24th International Conference on World Wide Web (Florence: ). 10.1145/2740908.2742839 [ CrossRef ] [ Google Scholar ]
  • Thelwall M. (2017). Microsoft Academic: a multidisciplinary comparison of citation counts with Scopus and Mendeley for 29 journals . J. Informet. 11 , 1201–1212. 10.1016/j.joi.2017.10.006 [ CrossRef ] [ Google Scholar ]
  • Thelwall M. (2018a). Can Microsoft Academic be used for citation analysis of preprint archives? The case of the social science research network . Scientometrics 115 , 913–928. 10.1007/s11192-018-2704-z [ CrossRef ] [ Google Scholar ]
  • Thelwall M. (2018b). Does Microsoft Academic find early citations . Scientometrics 114 , 325–334. 10.1007/s11192-017-2558-9 [ CrossRef ] [ Google Scholar ]
  • Thelwall M. (2018c). Microsoft Academic automatic document searches: accuracy for journal articles and suitability for citation analysis . J. Informet. 12 , 1–9. 10.1016/j.joi.2017.11.001 [ CrossRef ] [ Google Scholar ]
  • Traag V. A., Waltman L. (2019). Systematic analysis of agreement between metrics and peer review in the UK REF . Palgrave Commun. 5 :29. 10.1057/s41599-019-0233-x [ CrossRef ] [ Google Scholar ]
  • Turney P. D., Pantel P. (2010). From frequency to meaning: vector space models of semantics . J. Art. Intell. Res. 37 , 141–188. 10.1613/jair.2934 [ CrossRef ] [ Google Scholar ]
  • Walker D., Xie H., Yan K. K., Maslov S. (2007). Ranking scientific publications using a model of network traffic . J. Statist. Mech. 2007 :6010. 10.1088/1742-5468/2007/06/P06010 [ CrossRef ] [ Google Scholar ]
  • Waltman L., Eck N. J. V. (2012). The inconsistency of the h-index . J. Assoc. Informat. Sci. Technol. 63 , 406–415. 10.1002/asi.21678 [ CrossRef ] [ Google Scholar ]
  • Waltman L., Yan E. (2014). PageRank-related methods for analyzing citation networks, in Measuring Scholarly Impact , eds Waltman L, Yan E. (Cham: Springer; ), 83–100. 10.1007/978-3-319-10377-8_4 [ CrossRef ] [ Google Scholar ]
  • Wilhite W., Fong E. A. (2012). Coercive citation in academic publishing . Science 335 , 542–543. 10.1126/science.1212540 [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Wilsdon J. (2015). We need a measured approach to metrics . Nature 523 , 129–129. 10.1038/523129a [ PubMed ] [ CrossRef ] [ Google Scholar ]
  • Share full article

Advertisement

Supported by

Microsoft Debates What to Do With A.I. Lab in China

Amid U.S.-China tensions, the company has faced questions over whether operating an advanced research lab in Beijing is politically tenable.

Bill Gates and Xi Jinping shake hands on a large screen at the end of a long outdoor portico, photographed from below.

By Karen Weise ,  Cade Metz and David McCabe

Karen Weise reported from Seattle, Cade Metz from San Francisco and David McCabe from Washington.

When Microsoft opened an advanced research lab in Beijing in 1998, it was a time of optimism about technology and China.

The company hired hundreds of researchers for the lab, which pioneered Microsoft’s work in speech, image and facial recognition and the kind of artificial intelligence that later gave rise to online chatbots like ChatGPT . The Beijing operation eventually became one of the most important A.I. labs in the world. Bill Gates, Microsoft’s co-founder, called it an opportunity to tap China’s “deep pool of intellectual talent.”

Listen to This Article

Open this article in the New York Times Audio app on iOS.

But as tensions between the United States and China have mounted over which nation will lead the world’s technological future, Microsoft’s top leaders — including Satya Nadella, its chief executive, and Brad Smith, its president — have debated what to do with the prized lab for at least the past year, four current and former Microsoft employees said.

The company has faced questions from U.S. officials over whether maintaining a 200-person advanced technologies lab in China is tenable, the people said. Microsoft said it had instituted guardrails at the lab, restricting researchers from politically sensitive work.

The company, which is based in Redmond, Wash., said it had also opened an outpost of the lab in Vancouver, British Columbia, and would move some researchers from China to the location. The outpost is a backup if more researchers need to relocate, two people said. The idea of shutting down or moving the lab has come up, but Microsoft’s leaders support continuing it in China, four people said.

“We are as committed as ever to the lab and the world-class research of this team,” Peter Lee, who leads Microsoft Research, a network of eight labs across the world, said in a statement. Using the lab’s formal name, he added, “There has been no discussion or advocacy to close Microsoft Research Asia, and we look forward to continuing our research agenda.”

The debate at Microsoft stands out because the company is one of the few major U.S. tech firms — alongside Apple and Tesla — to keep a foothold in China. As China nurtured a domestic tech industry and geopolitical tensions increased with the United States, American companies such as Google whittled down their presence there. Facebook and other U.S. social media sites such as X have been blocked in China for years.

LinkedIn, which Microsoft owns, shut down its professional social network in China in 2021, citing growing compliance demands. But Microsoft has maintained its Bing search engine as the only foreign search engine in China, though it is heavily censored , and it offers its Windows operating system, cloud computing and applications for corporate customers there.

Microsoft has debated the lab’s future for several years, five people with knowledge of the situation said. It has become a target of national security concerns amid the rise of A.I. and growing aggression between the United States and China. The hypothetical risks are that China could hack or otherwise infiltrate the lab, or that its researchers could leave Microsoft to join Chinese companies that work closely with the government, the people said.

The Biden administration privately asked Microsoft about the lab while drafting a ban over the past two years on new U.S. investments in companies building sensitive technologies in China that Beijing could use to enhance its military, two people familiar with the conversations said. (The proposed rules, issued in August , are not yet final.)

Senators asked Mr. Smith about Microsoft’s ties to China at a subcommittee hearing on A.I. in September. He said the country accounted for 1.5 percent of Microsoft’s sales, which were $212 billion last fiscal year.

Microsoft faces “a tricky balance,” said Chris Miller, the author of “Chip War,” a book that tracks the geopolitical history of tech. “They need to consider where the trust of the political system is going.”

The White House declined to comment.

Microsoft’s lab in Beijing was born when Mr. Gates appointed Kai-Fu Lee, a Taiwanese-born A.I. researcher, to build the operation. (Dr. Lee later left to join Google and now runs a venture capital firm.)

Researchers at the lab, many of whom were at the top of their field, explored technologies such as speech recognition, computer vision and natural language understanding, which are cornerstones in the development of artificial intelligence. Some of the lab’s researchers left for key positions at Chinese tech giants like Alibaba, Baidu and Tencent or helped found start-ups such as Megvii, a facial recognition company that has contributed to a vast national surveillance system in the country.

In 2018, Microsoft said it had invested more than $1 billion in research and development in China over the previous decade. The technical talent and invention from the Beijing lab underpin a key internal argument for supporting it.

But the lab’s success and prestige also attracted attention in Washington, where the White House has increasingly restricted China’s access to crucial technologies, citing national security.

Microsoft’s leaders have discussed how to manage the tensions. Mr. Gates, who is still in regular touch with company executives and supports global engagement, has long backed the Beijing lab, people with knowledge of the matter said. He traveled to China in June and met President Xi Jinping, who told him that he was “the first American friend I’ve met with this year.”

Microsoft’s technology and research leaders, including Peter Lee and Kevin Scott, the chief technology officer, also support the lab, arguing that it has produced critical technological breakthroughs, two people said. Mr. Smith also backs the lab.

“The lesson of history is that countries succeed when they learn from the world,” Mr. Smith said in a statement. “Guardrails and controls are critical, while engagement remains vital.”

In recent years, Microsoft has limited what projects the researchers in China can work on, people with knowledge of the matter said. Last fall, researchers in China were not allowed on the small teams at Microsoft that had early access to GPT-4, the advanced A.I. system developed by Microsoft’s partner OpenAI, they said.

The lab also has restrictions on work related to quantum computing, facial recognition and synthetic media, Microsoft said. The company also blocks hiring or working with students and researchers from universities affiliated with China’s military, it said.

(The New York Times sued OpenAI and Microsoft last month for copyright infringement over the training of their A.I. systems.)

In the outpost of the lab in Vancouver, researchers can freely have access to critical technologies, including computing power and OpenAI systems needed for cutting-edge research, two people with knowledge of the lab said.

Kate Conger contributed reporting from San Francisco.

Audio produced by Sarah Diamond .

An earlier version of this article misstated the number of people working at a Microsoft research lab in Beijing. There are 200 people at the lab, not 800.

How we handle corrections

Karen Weise writes about technology and is based in Seattle. Her coverage focuses on Amazon and Microsoft, two of the most powerful companies in America. More about Karen Weise

Cade Metz writes about artificial intelligence, driverless cars, robotics, virtual reality and other emerging areas of technology. More about Cade Metz

David McCabe covers tech policy. He joined The Times from Axios in 2019. More about David McCabe

Cart

  • SUGGESTED TOPICS
  • The Magazine
  • Newsletters
  • Managing Yourself
  • Managing Teams
  • Work-life Balance
  • The Big Idea
  • Data & Visuals
  • Reading Lists
  • Case Selections
  • HBR Learning
  • Topic Feeds
  • Account Settings
  • Email Preferences

Why Microsoft Measures Employee Thriving, Not Engagement

  • Dawn Klinghoffer
  • Elizabeth McCune

microsoft research articles

And what other organizations can learn from their findings.

As the pandemic continues and many people work hybrid schedules, people analytics researchers at Microsoft realized they needed to move from measuring employee engagement to measuring employee thriving. Defined as “to be energized and empowered to do meaningful work,” the authors explain how their mindset shifted around this topic, what they’ve discovered among their employees, and what your organization can learn from their research.

One thing is clear: None of us are the same people today as we were prior to 2020. So, as our employees change, the ways we can best empower them need to evolve, too.

microsoft research articles

  • Dawn Klinghoffer is the head of people analytics at Microsoft.
  • Elizabeth McCune is the director of employee listening systems and culture measurement at Microsoft.

Partner Center

  • Work & Careers
  • Life & Arts

Become an FT subscriber

Limited time offer save up to 40% on standard digital.

  • Global news & analysis
  • Expert opinion
  • Special features
  • FirstFT newsletter
  • Videos & Podcasts
  • Android & iOS app
  • FT Edit app
  • 10 gift articles per month

Explore more offers.

Standard digital.

  • FT Digital Edition

Premium Digital

Print + premium digital.

Then $75 per month. Complete digital access to quality FT journalism on any device. Cancel anytime during your trial.

  • 10 additional gift articles per month
  • Global news & analysis
  • Exclusive FT analysis
  • Videos & Podcasts
  • FT App on Android & iOS
  • Everything in Standard Digital
  • Premium newsletters
  • Weekday Print Edition

Complete digital access to quality FT journalism with expert analysis from industry leaders. Pay a year upfront and save 20%.

  • Everything in Print
  • Everything in Premium Digital

The new FT Digital Edition: today’s FT, cover to cover on any device. This subscription does not include access to ft.com or the FT App.

Terms & Conditions apply

Explore our full range of subscriptions.

Why the ft.

See why over a million readers pay to read the Financial Times.

International Edition

  • Open supplemental data
  • Reference Manager
  • Simple TEXT file

People also looked at

Original research article, a review of microsoft academic services for science of science studies.

microsoft research articles

  • Microsoft Research, Redmond, WA, United States

Since the relaunch of Microsoft Academic Services (MAS) 4 years ago, scholarly communications have undergone dramatic changes: more ideas are being exchanged online, more authors are sharing their data, and more software tools used to make discoveries and reproduce the results are being distributed openly. The sheer amount of information available is overwhelming for individual humans to keep up and digest. In the meantime, artificial intelligence (AI) technologies have made great strides and the cost of computing has plummeted to the extent that it has become practical to employ intelligent agents to comprehensively collect and analyze scholarly communications. MAS is one such effort and this paper describes its recent progresses since the last disclosure. As there are plenty of independent studies affirming the effectiveness of MAS, this paper focuses on the use of three key AI technologies that underlies its prowess in capturing scholarly communications with adequate quality and broad coverage: (1) natural language understanding in extracting factoids from individual articles at the web scale, (2) knowledge assisted inference and reasoning in assembling the factoids into a knowledge graph, and (3) a reinforcement learning approach to assessing scholarly importance for entities participating in scholarly communications, called the saliency, that serves both as an analytic and a predictive metric in MAS. These elements enhance the capabilities of MAS in supporting the studies of science of science based on the GOTO principle, i.e., good and open data with transparent and objective methodologies. The current direction of development and how to access the regularly updated data and tools from MAS, including the knowledge graph, a REST API and a website, are also described.

Introduction

Centuries of scientific advancements have been a result of a virtuous cycle where scientists meticulously collect observation data to deduce a theoretical model and then use the model to predict new experimental outcomes as a means to validate the theory. This scientific principle has been applied to study the science of science, namely, the development of science itself, a field that sees pioneers like Eugene Garfield at the Institute for Scientific Information (ISI, now part of Clarivate Analytics) ( Garfield, 1955 , 1964 , 1972 ). Driven by the insights that scientific advancements inevitably leave abundant traces in the scholarly communications that often manifest themselves in the form of citations, a central topic in the science of science has been deriving quantitative models from citations for the purpose of analyzing and understanding the impacts of scientific work. Historically, citations made in the main body of an article have been difficult to collect so the bibliography has been used in their stead. Implicitly, this practice assumes the relations among publications can be approximated by the pairwise Boolean measures between the citing and the cited articles. Such an approximation is found to be too reductive in contrast to peer reviews for article-level assessments ( Wilsdon, 2015 ), although there is evidence suggesting noises in such a simplified model may be “canceled out” through aggregations at a level higher than individual articles ( Traag and Waltman, 2019 ). Indeed, the most widely used bibliometrics, such as the journal impact factor (JIF) or the h-index, are by design aggregate measures at the journal or the author level. However, the demands for article-level metrics are so strong that they make popular a practice assuming articles in the same journal are equal in quality and the aggregate metrics for the journal can serve as a proxy for the articles published therein. Its adverse effects are so profound and misuses so pervasive that renowned institutions and thought leaders have found it necessary to proclaim the San Francisco Declaration of Research Assessment (DORA) 1 to publicize a strong stance against using journal-level metrics for research assessments. A widely accepted good model to understand the impacts of individual publications has yet to be found.

Another challenge in the study of science of science is the explosive growth in the volume of scientific reports and the diversity of research topics. These have outstripped the cognitive capacity of human beings to properly digest and catch up. This cognitive overload ostensibly impacts everyone, including those employed by vendors to curate data and develop commercial platforms for science of science studies. As a result, errors and omissions in manually curated data are abundant, eroding the trustworthiness of studies conducted on those platforms. Most frustratingly, the proprietary and opaque nature in the commercial systems prevent recourses when obvious errors are spotted. As data-driven decision-making processes have become more prevalent in recent years, the platform quality has become a serious issue that prompts the Computing Research Association (CRA) to release a stern statement on the worsening state of commercial data and call for actions against unscientific practices based on or leading to flawed data 2 . In their report ( Berger et al., 2019 ), a CRA working group illustrates faulty data from Clarivate Analytics and surveys from humans no longer up to date in their fields may have led US News & World Report to produce abhorrent rankings on research universities that can severely mislead students in making school choices and funders in allocating resources. Similar to DORA, the CRA working group publishes a set of guidelines urging the adoption of Good and Open data with Transparent and Objective methodology, known as the GOTO principle, in conducting and publishing the results of quantitative science of science studies.

This article describes Microsoft Academic Services (MAS), a project in Microsoft Research with an aim to support researchers to follow the GOTO principle. Having evolved from the initially disclosed in ( Sinha et al., 2015 ), MAS now consists of three parts: an open dataset known as Microsoft Academic Graph (MAG) 3 , a freely available inference engine called Microsoft Academic Knowledge Exploration Service (MAKES), and a website called Microsoft Academic 4 that provides a more human friendly interface to MAKES. MAS is a part of an ongoing research that explores the nature of cognition, a topic in artificial intelligence (AI) that studies the mental capacity in acquiring, reasoning and inferencing with knowledge. The research is motivated by the observation that cognition involves the capabilities of memorizing, computing, being attentive, and staying focused on the task at hand, all of which can be programmed to the modern computer to outperform humans. Particularly for MAS, the project explores the boundary within which the power of machines can be harnessed to understand the scholarly communications observable on the web. In other words, MAS aims at developing AI agents that are well-read in all scientific fields and hopefully can become trustable consultants to human researchers on matters of scholarly activities taking place on the web. In this sense, the MAG component in MAS is the outcome of the knowledge acquisition and reasoning and MAKES, the capability of machine inferencing with the knowledge in MAG. The dataset MAG is distributed and frequently updated under an open data license and the inference algorithms in MAKES are published in relevant peer-review venues and summarized later in this article.

Aside from being open in data and transparent in algorithm as per the GOTO principle, MAS actively uses technologies to capture scholarly communication activities with adequate quality and coverage to strive for a good platform. To address the explosive growth in scientific research, MAS employs the state-of-the-art AI technologies, such as natural language understanding, to extract the knowledge from the text of these publications. This allows MAS to always take a data-driven approach in providing consistent data quality and avoid manual efforts that are often the source of subjective controversies or errors. Knowledge extraction in MAS goes beyond simply indexing key phrases to recognize and disambiguate the entities underpinning scholarly communications. MAS currently includes entities that describe who supported by which institutions have made what claims in which publication at which instance of which venue , as illustrated in Figure 1 . With more scholarly communications being conducted online with data and software tools, the definition of publication in MAS has been expanded. Aside from the traditional forms such as books, journals and conference papers, MAS has recognized datasets and software packages as additional forms of publications. Additionally, as plenty of scholarly work exerts impacts through commercial exploitation preceded by patent applications, MAS has also included them as publications. These new resources fit well into the model of publication entity in Figure 1 because they all have authors, affiliations, topical contents, etc., and can receive citations. In addition to extracting these entities, a key mission of knowledge extraction is to recognize the relations among the entities, such as the citation contexts characterizing how the work in one publication is received by others citing it. As schematized in Figure 1 , these entities and their relations are represented in a graph structure as the nodes and edges, respectively, leading to the name of MAG. Note that the entity recognition and disambiguation (ERD), as reported in ( Carmel et al., 2014 ), is far from a solved problem. However, the key here is the AI technologies employed in MAS are designed to learn and improve by itself by repeatedly reading more materials than any human can possibly do in a lifetime. After years of self-improving, many independent studies have suggested that MAG data are in many aspects as accurate, if not more, than manually curated data ( Herrmannova and Knoth, 2016 ; Harzing and Alakangas, 2017 ; Hug and Brändle, 2017 ; Hug et al., 2017 ; Thelwall, 2017 , 2018a , b , c ; Kousha et al., 2018 ).

www.frontiersin.org

Figure 1 . The data model of scholarly communications in MAS where the nodes represent the entity types modeled in MAG, and the simple and block arrows depict one-to-one and one-to-many relations among the entities, respectively.

Secondly, MAS uses technologies for scale, particularly when the lack of coverage in many datasets is becoming ever more concerning. While it might be appropriate in the last century for human experts to manually select only some of the scholarly communications into a database, this practice may have finally lived out its usefulness as the case studies in the CRA report have shown. Furthermore, with the advancements in information technology, online publishing has become a widely adopted medium for scientists to communicate with one another. Important activities, including self-archiving, data and software sharing, and community efforts dedicated to reproducing previously published results [e.g., Papers with Code 5 , ReScience ( Rougier et al., 2017 )] are taking place exclusively on the web. A modern dataset therefore must be able to capture all these web-only activities to properly reflect the current state of the reality, and it is hard to fathom how all these capturing efforts can be accomplished by hand. MAS provides an encouraging example that technologies can help in this area.

The key to MAS is large-scale deployment of AI agents in understanding scholarly communications. Therefore, the rest of the article is devoted to describing the methodologies so that the characteristics of MAS can be better understood. The AI technologies used in MAS, as illustrated in Figure 2 , encompass three areas: (1) natural language understanding, including ERD and concept detection to extract factoids from individual publication and to fulfill queries in MAKES, (2) knowledge reasoning to organize the factoids into MAG, and (3) a reinforcement learning system to learn a probabilistic measure called the saliency that facilitates the statistical learning and inferences in the above two areas.

www.frontiersin.org

Figure 2 . AI and service components in MAS are comprised of two feedback loops, one to grow the power of acquiring knowledge in MAG and the other to assess the saliency of each entity in MAG. In the first loop, each publication on the web is first processed by the MAG assisted entity recognition and disambiguation as described in (1). As the raw entities and their relations are extracted from individual publications, semantic reasoning algorithms are then applied to conflate them into a revised graph, including the concept hierarchy from all the publications. The revised MAG is then used in the next run to better extract entities from publication. The second loop utilizes the citation behaviors as the rewarding target for a reinforcement learning algorithm to assess the importance of each entity on MAG based on the network topology. The quantitative measure, called the saliency, serves as a ranking factor in MAKES, a search and recommendation engine for MAG.

Entity Recognition and Disambiguation

Central to MAS is the quest to harness the power of machine to acquire knowledge from written text. As alluded previously, the knowledge acquisition task amounts to recognizing the lexical constructs of the semantic objects representing either entities or relations. To be more precise, the task of natural language understanding in MAS is formulated as a maximum a posteriori (MAP) decision problem:

where the input x = ( w 1 , w 2 , ⋯ ) is a word sequence of a natural language expression, K is a knowledge base, and the task is to find the best output ŷ = ( e 1 , e 2 , ⋯), e i ∈ K , that is a sequence of semantic objects. For example, suppose the input is a sentence “HIV causes AIDS.” The ideal output should consist of two entities “HIV” and “AIDS,” and a relation “causing” between them.

The MAP decision is known to be optimal provided the posterior probability distribution in (1) can be accurately estimated. While this can be done directly, MAS uses a mathematically equivalent approach, known as the generative modeling, where the Bayes rule is applied to (1) to rewrite the MAP decision as:

with P ( x | y, K ) and P ( y | K ) the semantic language and the prior models, respectively. The semantic language model characterizes how frequently a sequence of semantic objects y is expressed through the word sequence x . Typically, an entity is lexicalized by a noun phrase while a relation, a verb phrase. MAS, however, does not utilize the syntax structure of natural language but, rather, just assumes that the lexical realization of each semantic object is statistically independent of one another, namely:

where x i denotes the i -th phrase segment in x corresponding to e i . Essentially, the semantic language model characterizes the synonymous expressions for each semantic object e i and how likely each of them is used. For example, the journal “Physical Review Letters” can be referred to by its full name, a common abbreviation “Phys Rev Lett,” or simply the acronym “PRL,” and an author can be mentioned using the last name, the first name or just its initial with an optional middle initial. The bibliography section, the text body and the web pages of a paper all provide abundant materials to harvest synonymous expressions. With large enough data samples, it appears adequate in MAS to use a simple maximum likelihood estimation, i.e., frequency counts with statistical smoothing, for the synonym model P (· | e i , K ).

The semantic prior model P ( y | K ) assesses the likelihood of a certain combination of semantic objects that can be derived from the knowledge base. In a way, the brunt of the statistical independent assumption in (3) is lessened because the contextual dependencies leading to a viable semantic interpretation are strictly enforced here. This can be seen by applying the chain rule of conditional probability to further decompose the semantic prior model as:

where P ( e 1 | K ) is the saliency of the entity e 1 and P ( e i | e i −1 , ⋯ e 1 , K ) is the semantic cohesion model according to the knowledge K . In conjunction with the synonym model, the semantic cohesion model can be estimated directly from data with an additional constraint that assigns zero probability to implausible semantic object combinations. This constraint plays a critical role in reducing the degree of ambiguities in understanding the input. For example, “Michael Evans” with a missing middle initial is a very confusable name, and “WWW” can mean a conference organized by IW3C2, a journal (ISSN: 1386-145X or 1573-1413), or even as a key word in the title of a paper. However, there are only two authors, a “Michael P. Evans” and a “Michael S. Evans” that have ever published any papers in the WWW conference, in 2002 and the other in 2017, respectively, and never in the namesake journal or any paper containing “WWW” as a key term in all other publication venues. If the publication year is also present, the apparently ambiguous input “Michael Evans ( Evans and Furnell, 2002 )” can be precisely resolved into the entity referring to the author named “Michael P. Evans” that has published a paper in “the eleventh International World Wide Web Conference” held in Honolulu Hawaii in the year of 2002. Using the knowledge-imposed constraints is particularly effective for author disambiguation when the technique is applied to understand curricula vitae or author homepages posted on the web. Assuming each such web page belongs to a single author, the publications listed therein are often high-quality signals to ascertain the identity of the author from the namesakes.

The manner that the knowledge is utilized in (4) also allows MAS to identify and acquire new synonymous expressions for existing entities and, often, new entities. This capability in acquiring new knowledge without human intervention is the key for MAS to enrich itself gradually. Mathematically, let K t denote the knowledge base used in (1) leading to the understanding of the scholarly materials y t at time t , the knowledge enrichment in MAS at an interval of Δ t , is also formulated as the MAP decision:

The iterative process of (5) in MAS can be better appreciated through a common task in parsing the bibliography where the author intent is to refer to a publication with a sequence of references to authors, followed by an optional publication title and a reference to a publication venue. The manners with which a reference is made, however, is highly inconsistent. When the semantic knowledge is applied to parse an input such as “Zhihong Shen, Hao Ma, and Kuansan Wang, ACL-2018,” it allows MAS to recognize fragments of the text, say, “Hao Ma” or “Kuansan Wang” as authors as they are frequently seen in the knowledge base. With these anchors, MAS can use (4) to infer that “Zhihong Shen” and “ACL-2018” are likely references to another author and the venue, respectively. These inferences can be made even before the publication records of ACL-2018 are included into the knowledge base and can be used with (5) to grow new entities in MAS.

While MAG only publishes the canonical expression for each entity, MAKES includes the probabilistic models that we derive from all the raw materials mentioned above. A step-by-step examination of (2) can be conducted in the query input box at the Microsoft Academic website where, upon each character entered, an API call is made into MAKES to analyze the semantic intent of the typed input with the MAP decision rule described in (2). Top interpretations manifesting themselves as query completions or suggestions are displayed to the user as a means for query intent, disambiguation or confirmation. More details are described in the FAQ page of the website 6 .

Concept Detection and Taxonomy Learning

Like many complex systems, the relations among entities in the scholarly communications ( Figure 1 ) cannot fully capture the activities because the semantics of the communications is encoded not in the topology but in the natural language contents of the publications. To address this issue, MAS adopts an entity type, called concepts [called “fields of study” in Sinha et al. (2015) ], to represent the semantic contents of a document. Unlike physical entities such as authors and affiliations, concepts are abstract and hence have no concrete way to define them. Furthermore, concepts are hierarchical in nature. For example, “machine learning” is a concept frequently associated with “artificial intelligence” that, in turn, is a branch of “computer science” but often intersects with “cognitive science” in “psychology.” Accordingly, a taxonomy must allow a concept to have multiple parents and organize all concepts into a directed acyclic graph (DAG). While concepts can be associated with all types of physical entities, say, to describe the topics of interest of a journal or the fields of expertise of a scholar, MAS only infers the relations between a publication and its concepts directly and leaves all others to be indirectly aggregated through publications.

A survey on the concepts taxonomy used in major library systems, presumably developed by human experts, suggests that few of them are compatible with each other. The low agreement among human experts leads MAS to create a concept taxonomy by itself solely from the document collection. As there are close to 1 million new publications a month being added in recent months, the machine learned taxonomy is dynamically adjusted on a regular basis so that new concepts can be added and obsolete concepts can be retired or merged with others.

Concept detection is a natural language understanding problem and, therefore, its mathematical foundation is also governed by (1). Unlike the ERD problem, however, the ideal output y ^ in this case is an unordered collection of DAGs of concepts rather than a sequence of semantic objects, and the textual boundaries of a concept in x are intrinsically soft, i.e., phrase segments can overlap. MAS therefore employs the approach to directly estimate the probabilistic distribution in (1) from the text rather than going through a generative model of (2). As detailed in the recent publication ( Shen et al., 2018 ), the key concept underlying the MAS approach here is the distributional similarity hypothesis proposed in 1950's ( Harris, 1954 ), which observes that semantically similar phrases tend to occur in similar contexts. There have been plenty of methods reported in the literature demonstrating the efficacy of applying distributional similarity for concept detection, either by training a hierarchical classifier mapping a sequence of discrete words directly into concepts, or by the embedding method that first converts the text into a vector representation with which learning and inferences can be conducted in a vector space ( Turney and Pantel, 2010 ). When properly executed, semantically similar phrases can be transformed into vectors close to one another, simplifying the synonymous expression detection, needed for (3), into a nearest neighbor search. In other words, the probabilistic distribution of synonyms P (·| e i , K ) can be estimated by the distance in the vector space. Recently, the embedding methods have produced many surprising results, starting with ( Mikolov et al., 2013 ; Berger et al., 2019 ), that contribute to a renaissance of the vector space model thanks to the availability of big data and powerful computational resources. The current practice in MAS, however, has found it more powerful to combine both the discrete and the vector space approaches into a mixture model for concept learning ( Shen et al., 2018 ).

The concept detection software in MAS has been released as part of the MAG distribution. The package, called Language Similarity 7 , provides a function with which the semantic similarity of two text paragraphs can be quantified using the embedding models trained from the publications in the corresponding MAG version. This function in turn serves as a mixture component for another function that, for any paragraph, returns a collection of top concepts detected in the paragraph that exceed a given threshold. Again, interested readers are referred to the recent article ( Shen et al., 2018 ) for technical details.

Network Semantics Reasoning

As MAS sources its materials from the web notorious for its uneven data qualities, duplicate, erroneous and missing information abounds. Critical to MAS is therefore a process, called conflation, that can reason over partial and noisy information to assemble the semantic objects extracted from individual documents into a cohesive knowledge graph. A key capability in conflation is to recognize and merge the same factoids while adjudicating any inconsistencies from multiple sources. Conflation therefore requires reasoning over the semantics of network topology and many of the techniques in MAS described in Sinha et al. (2015) are still in practice today.

Recently, a budding research area focuses on extending the notion of distributional similarity from its natural language root to the network environment. The postulation is straightforward: similar nodes tend to have similar types of edges connecting to similar nodes. Similar to the natural language use case of representing entities and relations as vectors, the goal of this approach is to transform the nodes and edges of a network into vectors so that reasonings with a network can be simplified and carried out in the vector space with algebraic mathematics. Network semantics, however, is more complicated than natural language whose contextual relations are single dimensional in nature: a phrase is either left or right to another. A network has a higher order topology because a node can simultaneously connect to a wide variety of others with edges representing distinctive relations. Citation network is a simple example where one paper can be cited by two others that also have a citation relation between them. Citation network is considered simple as it only has a single type of nodes, publication, and a single type of relation, citing. In reality, scholarly communications also involve people, organizations, locations, etc. that are best described by a heterogeneous network where multiple types of nodes are connected by multiple types of edges, making the notion of distributional similarity more sophisticated. The research in heterogenous network semantics reasoning, especially in its subfields of network and knowledge graph embedding, is ongoing and highly active.

MAS has been testing the network embedding techniques on related entity recommendation and found it essential for each entity to have multiple embeddings based on the types of relations involved in the inferences. In other words, embedding is sensitive to the sense defining similarity. For example, two institutions can be regarded as similar because their publications share a lot in common either in contents, in authorships, in venues, or are being cited together by same publications or authors. The multitude of senses of similarity leads to multiple sets of embeddings, of which results are included in MAG distributions. As the research in this area is still ongoing and the techniques by no means matured, MAS applications can achieve better results by combining the embedding and the discrete inference techniques. One such example is reported in a recent paper ( Kanakia et al., 2019 ) that describes the method behind the current related publication recommendation in MAS. The user studies in this application show the best system uses both the distance of the text embeddings and the frequency of being cited together.

Assessing Entity Importance With Saliency

As the MAP decision in (1) also drives MAKES to rank the results y in response to a query x , the entity prior P ( e | K ) in (4) is a critical component for MAKES. The way the entity prior is estimated determines in which sense the ranking is optimized. Ideally, the prior should be the importance the entity has been perceived by the scholarly community in general. Recently, a new area of research, lumped under the name altmetrics ( Piwowar, 2013 ), has been advocating that the searching, viewing, downloading, or endorsement activities in the social media for a publication should be included in estimating the importance of the scholarly work. Having monitored these activities for the past few years, we have found altmetrics a good indicator gauging how a publication has gained awareness in the social media. Although being known is a necessary step for being perceived as important, our observations cannot exclude the possibility that a publication is searched and viewed more because it is repeatedly mentioned in another highly regarded work, or authored by influential scholars or even just from reputable organizations. Based on our observations and concerns about altmetrics in the community (e.g., Cheung, 2013 ), the current focus in MAS is on exploiting the heterogeneity of scholarly communications mentioned above to estimate the entity prior by first computing the importance of a node relative to others of the same type and then weighting it by the importance of its entity type.

Saliency: An Eigencentrality Measure for Heterogeneous Dynamic Network

The eigenvector centrality measure, or simply eigencentrality, has been long known as a powerful method to assess the relative importance of nodes in a network ( Franceschet, 2011 ). Developed in the early twentieth century, eigencentrality measures the importance of a node relative to others by examining how strongly this node is referred to by other important nodes. Often normalized as a probabilistic measure, eigencentrality can be understood as a likelihood of a node being named as most important in a survey conducted on all members in the network. The method is made prominent by Google in its successful adaptation of eigencentrality for its PageRank algorithm: the PageRank of a webpage is measured by the proportional frequency of the incoming hyperlinks weighted by the PageRank of the respective sources. In a distinct contrast to simple citation counts, two important considerations in PageRank are the frequency of mentions in the citing article counts, and the importance of the citing source matters. Google has demonstrated that PageRank can be successfully used to assess the importance of each web document.

There are, however, two major challenges in using the eigencentrality as an article-level metric in general. First, the eigencentrality is mathematically well-defined only if the underlying network is well connected. This mathematical requirement is often not met in real-life, neither in the citation networks nor the web graph. To tackle this problem, Google introduced a “teleportation” mechanism in PageRank in which the connection between two web pages is only 85% dependent on the hyperlinks between them. The rest of the 15%, called the teleportation probability, is reserved for the assumption that all webpages are connected to each other intrinsically and uniformly. While the teleportation mechanism serves Google well, it is found to be fragile and implausible for the citation network ( Walker et al., 2007 ; Maslov and Redner, 2008 ): the ranking of scholarly publications is overly sensitive to the choice of the teleportation probability, and the best choice suggests scientists only follow the bibliography half the time, with the other half randomly discovering articles from the entire research literature following a uniform distribution. Many PageRank inspired studies, as recently reviewed in ( Waltman and Yan, 2014 ), have also made the same observation and proposed remedies utilizing the heterogeneity of the scholarly communication network. They mostly, however, are in an early exploratory stage as the manners in modeling the heterogeneous interactions still contain many heuristics needed to be further validated. Secondly, even the well-connected issue can be addressed through a heterogeneous model, another challenge, as pointed out by many (e.g., Walker et al., 2007 ), is how to avoid treating eigencentrality as a static measure so that the time differences in citations can be taken into account. It is undesirable to treat an article that receives the last citations long ago as equal to one that has just received the same amount of citations today because results without a proper temporal adjustment exhibit a favorable bias toward older publications that have more time to collect citations.

MAS attacks these two challenges with a unified framework called saliency based on the following considerations. First, to address the underlying network as changing in time, saliency is defined as the stochastic process characterizing the temporal evolution of the individual eigencentrality computed from a snapshot of the network. Without making assumptions on its form, the autoregressive moving-average (ARMA) process, mathematically known to be able to approximate a non-stationary distribution to any precision with enough orders, is used to model the temporal characteristics of saliency. Surprisingly for MAS, a simple first order autoregressive (AR) process seems sufficient for the model to reach an ergodic solution (to be shown below), suggesting that the endorsement power of a citation can be treated as simply as an exponential decay with a constant half-life interval. This finding is a validation of the observation first reported in ( Walker et al., 2007 ).

Secondly, to account for the heterogeneity of the network, MAS uses a mixture model in which the saliency of a publication is a weighted sum of the saliencies of the entities related to the publication. By considering the heterogeneity of scholarly communications, MAS allows one publication to be connected to another through shared authors, affiliations, publication venues and even concepts, effectively ensuring the well-connectedness requirement is met without introducing a random teleportation mechanism. Mathematically, let s x ( t ) denote the saliency vector of the entities of type x at time t , with x = p specifically for the publication, the heterogeneous mixture model coupled with an AR process leads to:

where Δ t is the interval between the two successive network snapshots are taken, w p, x the (non-negative) weight of a type x node on the publication, τ the time decaying factor in the AR process, and A p, x the adjacency matrix characterizing the connection strength between a publication to any entity of type x . Currently, MAS considers all nodes of types x ≠ p to have equal connection to the publication, e.g., given a publication all of its authors and affiliations are treated as contributed equally to the saliency of the publication. In the meantime for publications citing one another, A p, p is set proportional to the number of mentions in the text body of the citing article to the cited work.

As the heterogeneous model treats the saliency of a publication as the combined saliencies of all entities related to it, s p ( t ) is therefore a joint probabilistic distribution. Accordingly, the saliency of a non-publication entity can be obtained by marginalizing the joint distribution, i.e.,

where A x, p = [δ ij ] and

Again, the current MAS implementation does not address how the credit of a publication should be assigned unevenly to its authors based on the author order as (8) implies all authors have equal contributions, the side effect of which, however, is each institutions associated with a publication will receive its credit proportional to the number of authors affiliated with the institution. Ostensibly, a more sophisticated model than (8) can be used where, for instance, the author sequence can play a role in determining δ ij . MAS reports the author sequence as well as the affiliation sequence for authors with multiple affiliations, but has not yet used them for the purpose of computing saliency.

Estimating Saliency With Reinforcement Learning

To avoid making a strong assumption that the latent variables τ and w p, x are constant, MAS uses reinforcement learning (RL) to dynamically choose the best values based on the reinforcement signals streaming in through the observations. The choice is motivated by the fact that the RL technique is known to be effective in tackling the exploitation vs. exploration tradeoff, which in MAS means a balanced treatment between the older and newer publications or authors that have unequal time to collect their due recognitions. Often, the challenge of applying RL is the reinforcement signals are hard to obtain. This is fortunately not the case in MAS because approximately half a million new publications with tens of million citations are discovered every 2 weeks (= Δ t ), and these new observations provide ample materials to be reinforcement signals. Assuming the scholar communications are eventually just, namely, more important publications will receive higher citations in the long run ( NΔt, N ≫ 1), the goal of the RL in MAS is to maximize the agreement between the saliencies of today and the citations accumulated NΔt into the future. Currently, MAS uses the maximum mutual information (MMI) as the quantitative measurement for the agreement, namely, if c ( t ) denotes the vector of citation mention counts for all publication, the objective of the RL in MAS is to find:

where < ·, · > denotes the inner product. The choice of MMI allows (9) to be a convex function so that it can be iteratively solved with a quasi-Newton method. An off-the-shelve software implementing a L-BFGS algorithm is used in MAS.

It is a surprise that, by choosing long enough future N , the solutions to the latent variables τ and w p, x appear to be quite steady over time with the simplest form of ARMA process: 1 st order autoregression and no moving average. This apparent ergodicity allows MAS to administer the RL with a delay of NΔt ≈5 years, namely, the latent variables in (6) can be obtained by using the data observed up to 5 years ago to predict the citations of the recent 5 years. The results, as shown in Figure 3 , suggests that generally a publication accrues its saliency from citations at a weight slightly more than 92%, although the factors of its authors, affiliations, publication venues and even topics are non-trivial. Along the time domain, the value of τ, hovering around 0.9, corresponds to a temporal decay in saliency with a half-life of 7.5 years. In contrast to previously studies where the citations only account for 50% of the weight ( Walker et al., 2007 ; Maslov and Redner, 2008 ) or with a very short decay from 1 to 2.6 years ( Walker et al., 2007 ), the RL results in MAS are a much less dramatic departure from the common practice of using citation counts as an article level metric where, effectively, the metric is computed with a 100% weight in citations that do not decay over time.

www.frontiersin.org

Figure 3 . The longitudinal values of the latent variables underlying the saliency as obtained by the reinforcement learning (RL) algorithm. These latent variables correspond to the weighting the algorithm has to exert on each entity type in order to predict the future citation behaviors most optimally in the sense of Maximum Mutual Information, as described in (9). The model shows citations remain the dominant factor to have a high saliency. Despite a relatively simple configuration, the model exhibits remarkable stability over the 35 months shown in this figure other than the two instances, in April 2017 and July 2018, when MAG changed its treatment on affiliations dramatically.

As also shown in Figure 3 , the RL is not impervious to major changes in the underlying data, such as the treatments to author affiliations. In May 2017, a so-called “inferred affiliation” feature was introduced to MAG where authors with unknown affiliations were associated with the “most likely” institutions inferred from their recent publication records. An overly optimistic threshold led to many inaccurate projections, and the RL responded to the degradation in quality by lowering the affiliation weight and shifting it to citations. In July of the following year, the MAG data schema is altered to allow a single author to have multiple affiliations and all of which receive equal attribution from the publications by the author. Such a more faithful characterization of author affiliations leads to a boost of the affiliation weight from 1.5 to 3.5%, suggesting the RL mechanism finds the affiliation information more useful.

Properties of Saliency

The saliencies obtained with (6) and (7) are reported in MAG at each update interval for all entities in a quantized form of −1000ln s ( t ), and are used as the entity prior in the MAP decision (1) in MAKES that can be examined through the search and analytics results at Microsoft Academic websites. All these tools can be valuable for more and deeper investigations to fully understand the properties of saliency as a potential metric. For example, by design s p ( t ) further discriminates the following three citation behaviors not considered in the simple citation count: the number of mentions in the citing article, the age of the citations received, and the non-citation factors that can alleviate the disadvantages for newer publications. The combined effects of these three aspects on the article-level assessment can be further studied by inspecting the results from (1) with synthesized queries. Figure 4 shows a typical outcome of 20% disagreement in the ranking position differences between saliency and citation count based rankings using the query set ( Supplementary Material S1 ). A quick examination into the disagreements confirms that a publication can have a higher saliency, albeit lower citation counts, because it is cited by more prestigious or more recent work as designed. Where these disagreements are desired, however, is a question worth exploring.

www.frontiersin.org

Figure 4 . Histogram of ranking position by citation counts (CC) and saliencies (top) and their differences (bottom). Although future citation counts are the target for best estimating the saliencies, they only agree on the publication rankings roughly 80 percent of the time, demonstrating the effects of non-citation factors ( Figure 3 ) in the design of saliency. In contrast to citation counts, saliencies are sensitive to the venues, the authors, the concepts, and the recencies of the citing sources.

The design to unshackle the reliance on the overly reductive citation counts may also lead the saliency to be less susceptible to manipulations, ranging from citation coercions ( Wilhite and Fong, 2012 ) to malicious cheating ( López-Cózar et al., 2014 ) targeting metrics like the h-index. By using the citation contexts in saliencies, these manipulations are, in theory, less effective and easier to detect, as demonstrated by PageRank for the link spam detection in the web graph ( Gyöngyi and Garcia-Molina, 2005 ). The extent to which the gain of the eigenvector-based method can be transported from the web graph to the scholarly network, however, awaits further quantification.

Another research topic MAS can be useful is in the effectiveness of saliencies of non-publication entities that, as described in (7), are aggregated from publication saliencies. This design gives rise to at least two intriguing properties. First, an entity can achieve high saliency with lots of publications, not all of which are important. As a result, saliency appears to be measuring both productivity and impact simultaneously, just like h-index. Indeed, shows a comparison between the h-index and the saliency of Microsoft authors ( Supplementary Material S2 ). Overall, there is a trend line suggesting individuals with a higher h-index tend to also have a higher saliency, but notable disagreements between the two abound. The author with the highest h-index, 134, in this set has the most publications at 619 articles that in total receive 60,157 citations but ranks only at the 4th place by saliency. Conversely, the highest ranked author by saliency has published only 138 papers receiving 82,293 citations with a h-index at 76. Most notably, the second highest ranked author by saliency has an h-index only at 31. This is because the author has published only 39 papers, which limits the h-index, but they are all well received with a total citation count of 58,268, which buoys the saliency. The drawbacks of the h-index, e.g., capping at the publication counts in this example, are well-known ( Waltman and Eck, 2012 ). By considering more factors and not limited to overly reductive raw signals, saliency appears to be better equipped to avoid mischaracterizing researchers who strive for the quality and not the quantity of their publications.

Secondly, because the underlying foundation of an aggregated saliency is based on article-level analysis, interdisciplinary work seems to be better captured. One such example is the journal ranking on a given subject, say, Library Science. As shown for a while at Microsoft Academic website 8 , journals like Nature and Science are among the top 10 for this field when ranked by saliency. This may be a surprise to many human experts because these two journals are seldom considered as a publication venue for the field of library science. Indeed, if the journals are ranked by h-index, these two journals will appear in much lower positions because the numbers of articles in the field are lower in these two journals. However, a closer investigation shows that these two journals have influential articles in the field, such as the Leiden Manifesto ( Hicks et al., 2015 ) in Nature and the coercive citation studies ( Wilhite and Fong, 2012 ) in Science. If one were to understand the most impactful papers in this field, precluding these two journals into consideration would lead to unacceptable omissions and result in incomplete work. Again, this example highlights the known problem of using journals as the unit to conduct quantitative scientific studies, and the sharp focus into article-level analysis, as demonstrated feasible by saliency, appears to be a better option.

Prestige: Size-Normalized Saliency

A known issue existing in aggregate measurements is that the sheer number of data points being considered can often play an outsized role. This can be seen in Figure 5 where the author saliency largely agrees, especially for prolific authors, with the h-index, a metric designed to measure the impact as well as the productivity. As implied by (7), an author can reach a high saliency by having a large number of publications despite most of them receive only moderate recognitions. Given it has been observed that hyper-prolific authors exist ( Ioannidis et al., 2018 ), and their publications seem to yield uneven qualities ( Bornmann and Tekles, 2019 ), it might be helpful to juxtapose the saliency with a corresponding size-normalized version, which we call prestige, to further discern the two aspects. To be specific, the prestige of a non-publication entity can be derived from (7) as:

where A ¯ x , p = [ δ i j / ∑ j δ i j ] in contrast to (9). In short, the prestige of an entity is the average of the saliencies of its publications. Figure 6 illustrates the effect of the size normalization through the rankings of the world research institutions in the field of computer science based on the saliencies and the prestiges of their research papers published during the 5-year window between 2012 and 2016. The institutions that publish with consistent impacts are lined up along the main diagonal where the size normalization has negligible effect on their rankings. It appears the majority of the institutions are in this category. Scattered to the upper left of the diagonal, however, are those that are not the most prolific institutions but, when they do, their publications tend to be highly recognized by the research community. Size normalization, as expected, significantly boosts their rankings, for instance, as in the casesof Princeton University and Google. On the other hand, clustered to the lower right to the diagonal are the institutions that achieve high saliencies by publishing a large body of literature, as reflected in the relatively large bubble sizes in Figure 6 , and hence their rankings are negatively impacted by the size normalization operation.

www.frontiersin.org

Figure 5 . A scatter plot comparing the h-index and the saliency where each dot corresponds to the h-index and the saliency of a Microsoft author. Although the two metrics largely agree, the saliency measure is able to overcome a known limitation of the h-index and highlight authors who have published widely recognized work but not in large quantity.

www.frontiersin.org

Figure 6 . Saliency (horizontal) vs. Prestige ranking of the top 50 research institutions in computer science area. The size of each bubble corresponds to the number of publications included in computing the saliency and the prestige measurements.

As many GOTO-compliant ranking systems ( Berger et al., 2019 ) have discovered, one cannot over-emphasize that institution rankings are a highly sophisticated task that necessitates multiple perspectives and with varying degrees in granularities that commercial rankings such as US News & World Report are typically ill-equipped. To illustrate the point, Figure 7 shows the saliency-prestige rankings of institutions in the subfield of computer science, artificial intelligence, and its subfields of machine learning, computer vision and natural language processing. The recurring themes emerging from the high variances in the ranking results and significant differences in the top institutions strongly suggest that the ranking result of a field is a very poor predictor of its subfields. This is consistent with our observation that, within the subfields of computer science, the spectrum of research topics is so broad that institutions can choose to specialize into a selective few to have a strong and highly impactful research program. Consequently, ranking institutions at too broad a category amounts to comparing research on notably different fields that can have distinct publication culture and citation behaviors, i.e., is an apple-vs.-orange type of comparison. With new resources like MAS that can pinpoint each publication to very fine-grained fields of study, such a deeply-flawed methodology that was previously tolerated due to data scarcity should no longer be deemed acceptable and must be soundly rejected by the community.

www.frontiersin.org

Figure 7 . Institution Rankings, by Saliency (horizontal) vs. Prestige, for the field of Artificial Intelligence and its subfields.

The explosive growth in scholarly communications has made it more difficult for individual humans to keep track of the latest achievements and trends in scientific research. The warning signs are visible in the worsening qualities of research assessments involving expert opinions, as a recent CRA study showed. This article describes how MAS utilizes the advancements in AI to curate a good and open data set and enable transparent and objective methodologies (GOTO) for scientific studies on science. The AI components in MAS, in natural language understanding, in knowledge reasoning and inferences, and in reinforcement learning for estimating saliencies of entities in scholarly communications have been described. There are early indications that saliencies, an objective measure by harvesting the peer reviewed citation contexts, avoid many drawbacks of existing academic metrics.

Data Availability Statement

All datasets generated for this study are included in the article/ Supplementary Material .

Author Contributions

KW drafted the manuscript and coordinated the research project. ZS, RR, and DE supervised the MAG, MAS, and the MAKES portions of the work. CH reviewed the experimental setups and software, while the rest of the authors have equal contributions to the data collected in the work.

The authors declare that this study received funding from Microsoft Research. The funder was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

Conflict of Interest

All authors are employed by the company Microsoft Research.

Acknowledgments

Dr. Hao Ma led the efforts in creating many advanced features in MAG. Dr. Bo-June Paul Hsu led the team to develop the inference engine in MAKES, and, with the assistance of Dr. Rong Xiao, implemented the first version of the reinforcement learning to compute saliency. The work would not be possible without the strong support from Microsoft Bing engineering teams and colleagues in Microsoft Research labs around the globe.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fdata.2019.00045/full#supplementary-material

Supplementary Material S1. A randomly synthesized query sets to study the differences between the citation count based and the saliency based ranking behaviors.

Supplementary Material S2. Analytical script and the data to study the difference between the h-index and saliencies for authors.

1. ^ https://sfdora.org/

2. ^ See https://cra.org/cra-statement-us-news-world-report-rankings-computer-science-universities/

3. ^ https://docs.microsoft.com/en-us/academic-services/

4. ^ https://academic.microsoft.com

5. ^ https://paperswithcode.com/

6. ^ https://academic.microsoft.com/faq

7. ^ https://www.microsoft.com/en-us/research/project/academic/articles/understanding-documents-by-using-semantics/

8. ^ See the analytic page at https://academic.microsoft.com/journals/41008148,161191863

Berger, E., Blackburn, S. M., Brodley, C., Jagadish, H. V., McKinley, K. S., Nascimento, M. A., et al. (2019). GOTO rankings considered helpful. Commun. ACM 62, 29–30. doi: 10.1145/3332803

CrossRef Full Text | Google Scholar

Bornmann, L., and Tekles, A. (2019). Productivity does not equal usefulness. Scientometrics 118, 705–707. doi: 10.1007/s11192-018-2982-5

Carmel, D., Chang, M. W., Gabrilovich, E., Hsu, B. J. P., and Wang, K. (2014). “ERD'14: entity recognition and disambiguation challenge,” in Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval (Gold Coast, QLD). doi: 10.1145/2600428.2600734

Cheung, M. K. (2013). Altmetrics: too soon for use in assessment. Nature 494, 176–176. doi: 10.1038/494176d

PubMed Abstract | CrossRef Full Text | Google Scholar

Evans, M., and Furnell, S. (2002). “A web-based resource migration protocol using WebDav,” in Proceedings of the WWW-2002 (Honolulu, HI).

Google Scholar

Franceschet, M. (2011). PageRank: standing on the shoulders of giants. Commun. ACM 54, 92–101. doi: 10.1145/1953122.1953146

Garfield, E. (1955). Citation indexes for science: a new dimension in documentation through association of ideas. Science 122, 108–111. doi: 10.1126/science.122.3159.108

Garfield, E. (1964). Science citation index- a new dimension in indexing. Science 144, 649–654. doi: 10.1126/science.144.3619.649

Garfield, E. (1972). Citation analysis as a tool in journal evaluation journals can be ranked by frequency and impact of citations for science policy studies. Science 178, 471–479. doi: 10.1126/science.178.4060.471

Gyöngyi, Z., and Garcia-Molina, H. (2005). Web Spam Taxonomy . Chiba: AIRWeb.

Harris, Z. S. (1954). Distributional structure. WORD 10, 146–162. doi: 10.1080/00437956.1954.11659520

Harzing, A. W., and Alakangas, S. (2017). Microsoft Academic: is the phoenix getting wings? Scientometrics 110, 371–383. doi: 10.1007/s11192-016-2185-x

Herrmannova, D., and Knoth, P. (2016). An Analysis of the Microsoft Academic Graph, D-lib Magazine 22, 6. doi: 10.1045/september2016-herrmannova

Hicks, D., Wouters, P., Waltman, L., Rijcke, S. D., and Rafols, I. (2015). Bibliometrics: the leiden manifesto for research metrics. Nature 520, 429–431. doi: 10.1038/520429a

Hug, S. E., and Brändle, M. P. (2017). The coverage of Microsoft academic: analyzing the publication output of a university. Scientometrics 113, 1551–1571. doi: 10.1007/s11192-017-2535-3

Hug, S. E., Ochsner, M., and Brändle, M. P. (2017). Citation analysis with microsoft academic. Scientometrics 111, 371–378. doi: 10.1007/s11192-017-2247-8

Ioannidis, J. P. A., Klavans, R., and Boyack, K. W. (2018). Thousands of scientists publish a paper every five days. Nature 561, 167–169. doi: 10.1038/d41586-018-06185-8

Kanakia, S. Z., Eide, D., and Wang, K. (2019). “A scalable hybrid research paper recommender system for microsoft academic,” in WWW '19 The World Wide Web Conference (New York, NY: ACM). doi: 10.1145/3308558.3313700

Kousha, K., Thelwall, M., and Abdoli, M. (2018). Can Microsoft Academic assess the early citation impact of in-press articles? A multi-discipline exploratory analysis. J. Informet. 12, 287–298. doi: 10.1016/j.joi.2018.01.009

López-Cózar, E. D., Robinson-García, N., and Torres-Salinas, D. (2014). The Google scholar experiment: How to index false papers and manipulate bibliometric indicators. J. Assoc. Inform. Sci. Technol. 65, 446–454. doi: 10.1002/asi.23056

Maslov, S., and Redner, S. (2008). Promise and pitfalls of extending Google's pagerank algorithm to citation networks. J. Neurosci. 28, 11103–11105. doi: 10.1523/JNEUROSCI.0002-08.2008

Mikolov, T., Sutskever, I., Chen, K., Corrado, G. S., and Dean, J. (2013). “Distributed representations of words and phrases and their compositionality,”in NIPS'13 Proceedings of the 26th International Conference on Neural Information Processing Systems , Vol. 2 (Lake Tahoe, NV), 3111–3119.

Piwowar, H. (2013). Altmetrics: Value all research products. Nature 493, 159–159. doi: 10.1038/493159a

Rougier, N. P., Hinsen, K., Alexandre, F., Arildsen, T., Barba, L. A., Benureau, F. C. Y., et al. (2017). Sustainable computational science: the ReScience initiative. PeerJ 3, 1–8. doi: 10.7717/peerj-cs.142

Shen, Z., Ma, H., and Wang, K. (2018). “A web-scale system for scientific knowledge exploration,” in Meeting of the Association for Computational linguistics (Melbourne, VIC), 87–92. doi: 10.18653/v1/P18-4015

Sinha, A., Shen, Z., Song, Y., Ma, H., Eide, D., Hsu, B. J. P., and Wang, K. (2015). “An overview of Microsoft Academic Service (MAS) and applications,” in Proceedings of the 24th International Conference on World Wide Web (Florence). doi: 10.1145/2740908.2742839

Thelwall, M. (2017). Microsoft Academic: a multidisciplinary comparison of citation counts with Scopus and Mendeley for 29 journals. J. Informet. 11, 1201–1212. doi: 10.1016/j.joi.2017.10.006

Thelwall, M. (2018a). Can Microsoft Academic be used for citation analysis of preprint archives? The case of the social science research network. Scientometrics 115, 913–928. doi: 10.1007/s11192-018-2704-z

Thelwall, M. (2018b). Does Microsoft Academic find early citations. Scientometrics 114, 325–334. doi: 10.1007/s11192-017-2558-9

Thelwall, M. (2018c). Microsoft Academic automatic document searches: accuracy for journal articles and suitability for citation analysis. J. Informet. 12, 1–9. doi: 10.1016/j.joi.2017.11.001

Traag, V. A., and Waltman, L. (2019). Systematic analysis of agreement between metrics and peer review in the UK REF. Palgrave Commun. 5:29. doi: 10.1057/s41599-019-0233-x

Turney, P. D., and Pantel, P. (2010). From frequency to meaning: vector space models of semantics. J. Art. Intell. Res. 37, 141–188. doi: 10.1613/jair.2934

Walker, D., Xie, H., Yan, K. K., and Maslov, S. (2007). Ranking scientific publications using a model of network traffic. J. Statist. Mech. 2007:6010. doi: 10.1088/1742-5468/2007/06/P06010

Waltman, L., and Eck, N. J. V. (2012). The inconsistency of the h-index. J. Assoc. Informat. Sci. Technol. 63, 406–415. doi: 10.1002/asi.21678

Waltman, L., and Yan, E. (2014). “PageRank-related methods for analyzing citation networks,” in Measuring Scholarly Impact , eds L, Waltman and E. Yan (Cham: Springer), 83–100. doi: 10.1007/978-3-319-10377-8_4

Wilhite, W., and Fong, E. A. (2012). Coercive citation in academic publishing. Science 335, 542–543. doi: 10.1126/science.1212540

Wilsdon, J. (2015). We need a measured approach to metrics. Nature 523, 129–129. doi: 10.1038/523129a

Keywords: microsoft academic services, microsoft academic graph, knowledge graph (KG), machine cognition, academic search, artificail intelligence (AI)

Citation: Wang K, Shen Z, Huang C, Wu C-H, Eide D, Dong Y, Qian J, Kanakia A, Chen A and Rogahn R (2019) A Review of Microsoft Academic Services for Science of Science Studies. Front. Big Data 2:45. doi: 10.3389/fdata.2019.00045

Received: 28 August 2019; Accepted: 18 November 2019; Published: 03 December 2019.

Reviewed by:

Copyright © 2019 Wang, Shen, Huang, Wu, Eide, Dong, Qian, Kanakia, Chen and Rogahn. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY) . The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Kuansan Wang, kuansanw@microsoft.com

Exclusive-Microsoft to Separate Teams and Office Globally Amid Antitrust Scrutiny

Reuters

FILE PHOTO: Microsoft Teams app is seen on the smartphone placed on the keyboard in this illustration taken, July 26, 2021. REUTERS/Dado Ruvic/Illustration/File Photo

By Foo Yun Chee

BRUSSELS (Reuters) - Microsoft will sell its chat and video app Teams separately from its Office product globally, the U.S. tech giant said on Monday, six months after it unbundled the two products in Europe in a bid to avert a possible EU antitrust fine.

The European Commission has been investigating Microsoft's tying of Office and Teams since a 2020 complaint by Salesforce-owned competing workspace messaging app Slack.

Teams, which was added to Office 365 in 2017 for free, subsequently replaced Skype for Business and became popular during the pandemic due in part to its video conferencing.

Rivals, however, said packaging the products together gives Microsoft an unfair advantage. The company started selling the two products separately in the EU and Switzerland on Aug. 31 last year.

"To ensure clarity for our customers, we are extending the steps we took last year to unbundle Teams from M365 and O365 in the European Economic Area and Switzerland to customers globally," a Microsoft spokesperson said.

"Doing so also addresses feedback from the European Commission by providing multinational companies more flexibility when they want to standardise their purchasing across geographies."

Microsoft said in a blogpost that it was introducing a new lineup of commercial Microsoft 365 and Office 365 suites that do not include Teams in regions outside the EEA (European Economic Area) and Switzerland, and also a new standalone Teams offering for Enterprise customers in those regions.

Starting April 1, customers can either continue with their current licensing deal, renew, update or switch to the new offers.

For new commercial customers, prices for Office without Teams range from $7.75 to $54.75 depending on the product while Teams Standalone will cost $5.25. The figures may vary by country and currency. The company did not disclose prices for current packaged products.

Microsoft's unbundling may not be enough to stave off EU antitrust charges which will likely be sent to the company in the coming months as rivals criticise the level of the fees and the ability of their messaging services to function with Office Web Applications in their own services, sources said.

Microsoft, which has racked up 2.2 billion euros ($2.4 billion) in EU antitrust fines in the past decade for tying or bundling two or more products together, risks a fine of as much as 10% of its global annual turnover if found guilty of antitrust breaches.

($1 = 0.9265 euros)

(Reporting by Foo Yun Chee; editing by Jason Neely)

Copyright 2024 Thomson Reuters .

Tags: United States , European Union , Europe

The Best Financial Tools for You

Credit Cards

microsoft research articles

Personal Loans

microsoft research articles

Comparative assessments and other editorial opinions are those of U.S. News and have not been previously reviewed, approved or endorsed by any other entities, such as banks, credit card issuers or travel companies. The content on this page is accurate as of the posting date; however, some of our partner offers may have expired.

microsoft research articles

Subscribe to our daily newsletter to get investing advice, rankings and stock market news.

See a newsletter example .

You May Also Like

Best reit etfs to buy and hold for 2024.

Glenn Fydenkevez March 29, 2024

microsoft research articles

9 Highest Dividend Stocks in the S&P 500

Jeff Reeves March 29, 2024

microsoft research articles

5 of the Best Stocks to Buy Now

Ian Bezek March 29, 2024

microsoft research articles

Handling Extra Portfolio Risk

Kate Stalter March 29, 2024

microsoft research articles

7 Best Copper Stocks to Buy

Glenn Fydenkevez March 28, 2024

microsoft research articles

11 New Spot Bitcoin ETFs to Buy in 2024

Tony Dong March 28, 2024

microsoft research articles

Is M&A Activity Set to Spike in 2024?

Scott Ward March 28, 2024

microsoft research articles

7 Oversold Tech Stocks to Buy

Wayne Duggan March 28, 2024

microsoft research articles

7 Top Gene-Editing Stocks to Buy

Wayne Duggan March 27, 2024

microsoft research articles

7 Best Preferred Stock ETFs to Buy Now

Tony Dong March 27, 2024

microsoft research articles

Fixed Index Annuity Is Making a Comeback

Marguerita Cheng March 27, 2024

microsoft research articles

9 Growth Stocks for the Next 10 Years

Jeff Reeves March 27, 2024

microsoft research articles

5 Best Short-Term Investments for Income

Tony Dong March 26, 2024

microsoft research articles

10 Best Health Care Stocks to Buy

Wayne Duggan March 26, 2024

microsoft research articles

Vanguard Funds for Beginners

microsoft research articles

Best Semiconductor Stocks

Brian O'Connell March 25, 2024

microsoft research articles

Best Long Term Stocks to Buy and Hold

Glenn Fydenkevez March 25, 2024

microsoft research articles

Best Investing Apps for Beginners

Kate Stalter March 25, 2024

microsoft research articles

10 Best Growth Stocks to Buy for 2024

Wayne Duggan March 25, 2024

microsoft research articles

Best Renewable Energy Stocks to Buy

Matt Whittaker March 22, 2024

microsoft research articles

microsoft research articles

IMAGES

  1. Microsoft Research on LinkedIn: This year Microsoft Research launched

    microsoft research articles

  2. Microsoft Research AI, The Lab To Develop General-Purpose AI Technology

    microsoft research articles

  3. PPT

    microsoft research articles

  4. Seven Microsoft Research researchers named ACM 2016 Fellows

    microsoft research articles

  5. Take a sneak peek at the Microsoft's AI research facility

    microsoft research articles

  6. How to Write a Research Article

    microsoft research articles

COMMENTS

  1. Publications index

    Publications index. Below is an index of publications written by Microsoft researchers, often in collaboration with the academic community.

  2. About Microsoft Research

    Researchculture. We are rigorous and objective: since its founding in 1991, Microsoft Research has been committed to an academic research approach that advances our understanding of the world and how we use technology to interact with it, and with each other. We take calculated risks: as an industrial research lab, we contribute ideas that help ...

  3. Microsoft Research

    Microsoft Research (MSR) is the research subsidiary of Microsoft.It was created in 1991 by Richard Rashid, Bill Gates and Nathan Myhrvold with the intent to advance state-of-the-art computing and solve difficult world problems through technological innovation in collaboration with academic, government, and industry researchers. The Microsoft Research team has more than 1,000 computer ...

  4. Microsoft study: Small and medium-sized businesses growing fast by

    Analysys Mason was commissioned to conduct research with over 3,000 SMBs from 10 different areas of the world, representing the global footprint of SMBs. Today, we are releasing our findings with the Microsoft SMB Voice and Attitudes to Technology Study 2022. Top insights. Digital technology adoption contributes to faster growth

  5. How Microsoft Became Innovative Again

    February 20, 2023. Smith Collection/Gado/Getty Images. Summary. How did Microsoft revive its culture of innovation? For years, the company has been written off for playing defense on its position ...

  6. Microsoft and LinkedIn share latest data and innovation for hybrid work

    Because in uncharted territory, we need to be able to shift and adjust as data and research offer new insights to guide our way. The Hybrid Work Paradox and the 'Great Reshuffle' A report out today on our Work Trend Index shares what we're learning from Microsoft employees in over 100 countries around the world. Employee surveys tell us ...

  7. A Review of Microsoft Academic Services for Science of Science Studies

    A scalable hybrid research paper recommender system for microsoft academic, in WWW '19 The World Wide Web Conference (New York, NY: ACM; ). 10.1145/3308558.3313700 [Google Scholar] Kousha K., Thelwall M., Abdoli M. (2018). Can Microsoft Academic assess the early citation impact of in-press articles? A multi-discipline exploratory analysis. J.

  8. Microsoft Debates What to Do With A.I. Lab in China

    An earlier version of this article misstated the number of people working at a Microsoft research lab in Beijing. There are 200 people at the lab, not 800. How we handle corrections.

  9. Why Microsoft Measures Employee Thriving, Not Engagement

    Summary. As the pandemic continues and many people work hybrid schedules, people analytics researchers at Microsoft realized they needed to move from measuring employee engagement to measuring ...

  10. Microsoft Research

    At Microsoft Research, we accelerate scientific discovery and technology innovation to empower every person and organization on the planet to achieve more. W...

  11. Getting Started with Microsoft Academic

    It is a free public academic search engine for publications and literature, developed by Microsoft Research. Fig. 1: Microsoft Academics Timelines since 2006. At the time of writing this article ...

  12. Microsoft Research

    Microsoft Research Forum returns on March 5 at 9:00 AM PT. Join us for Episode 2 of engaging discussions with the global research community on how AI is transforming the way we all live, work, and ...

  13. How Microsoft's Stumbles Led to Its OpenAI Alliance

    For more than a decade, Microsoft Research, the company's in-house research group, has touted artificial intelligence breakthroughs such as translating speech to text and software that could understand human language or recognize objects in images. But the company's effort to commercialize its AI research moved at more of a crawl, according to three employees involved in it.

  14. Microsoft names DeepMind co-founder Mustafa Suleyman as CEO of ...

    Hayden Field @haydenfield. Key Points. Mustafa Suleyman is a co-founder of DeepMind, which became a leading artificial intelligence research lab after Google bought it in 2014. Suleyman is ...

  15. How Big Tech is winning the AI talent war

    Microsoft's agreement with two founders of Inflection is the latest sign that deep-pocketed companies are scooping up much of the expertise in AI. ... The research company Zeki, which tracks the ...

  16. Microsoft and OpenAI are planning a massive $100 billion AI

    In the short term, Microsoft already approved a $1 billion data center expansion in Mount Pleasant, Wisconsin, that could grow into a $10 billion "Phase 4" system expected to go live in 2026.

  17. Frontiers

    Microsoft Research, Redmond, WA, United States; Since the relaunch of Microsoft Academic Services (MAS) 4 years ago, scholarly communications have undergone dramatic changes: more ideas are being exchanged online, more authors are sharing their data, and more software tools used to make discoveries and reproduce the results are being distributed openly.

  18. Exclusive-Microsoft to Separate Teams and Office Globally Amid

    Microsoft, which has racked up 2.2 billion euros ($2.4 billion) in EU antitrust fines in the past decade for tying or bundling two or more products together, risks a fine of as much as 10% of its ...

  19. Microsoft Corp: An Exploration into Its Intrinsic Value

    In this article, we will take a look into Microsoft Corp's (NASDAQ:MSFT) DCF analysis, a reliable and data-driven approach to estimating its intrinsic value. Instead of using future free cash flow ...

  20. Large Language Models Answer Medical Questions ...

    The core asset of scientific research - universal reproducibility - must be guaranteed and protected by clear and shared guidelines that strictly regulate use of LLMs. The self-learning procedures and related AIXIs must be tested for discrepancy with established and published clinical and economic data. This or other measures of accuracy and ...