Media Analytics News
https://www.nec-labs.com/wp-content/uploads/2023/11/The-Industrial-Lab-Advantage-Delivering-on-the-Expectations-of-AI.jpg 628 1200 NEC Labs America https://neclabs.wpengine.com/wp-content/uploads/2022/08/NEC-Labs-Blue-Logo-Square-300x267.jpg NEC Labs America2023-11-10 12:15:032023-11-10 12:45:04The Industrial Lab Advantage: Delivering on the Expectations of AI | Future of Work News
Our Chris White outlines how Invisible AI transforms our lives and its potential to bring about transformative societal changes, including safer space travel, reliable cell networks, smarter cities, productive factories, and efficient homes. Invisible AI will intelligently anticipate needs, automate tasks, and enhance the human experience.
https://neclabs.wpengine.com/wp-content/uploads/2022/08/NEC-Labs-Blue-Logo-Square-300x267.jpg 0 0 NEC Labs America https://neclabs.wpengine.com/wp-content/uploads/2022/08/NEC-Labs-Blue-Logo-Square-300x267.jpg NEC Labs America2023-10-25 00:00:002023-11-30 09:09:04Adaptation Speed Analysis for Fairness-Aware Causal Models
For example, in machine translation tasks, to achieve bidirectional translation between two languages, the source corpus is often used as the target corpus, which involves the training of two models with opposite directions. The question of which one can adapt most quickly to a domain shift is of significant importance in many fields. Specifically, consider an original distribution p that changes due to an unknown intervention, resulting in a modified distribution p*. In aligning p with p*, several factors can affect the adaptation rate, including the causal dependencies between variables in p. In real-life scenarios, however, we have to consider the fairness of the training process, and it is particularly crucial to involve a sensitive variable (bias) present between a cause and an effect variable. To explore this scenario, we examine a simple structural causal model (SCM) with a cause-bias-effect structure, where variable A acts as a sensitive variable between cause (X) and effect (Y). The two models respectively exhibit consistent and contrary cause-effect directions in the cause-bias-effect SCM. After conducting unknown interventions on variables within the SCM, we can simulate some kinds of domain shifts for analysis. We then compare the adaptation speeds of two models across four shift scenarios. Additionally, we prove the connection between the adaptation speeds of the two models across all interventions.
https://neclabs.wpengine.com/wp-content/uploads/2022/08/NEC-Labs-Blue-Logo-Square-300x267.jpg 0 0 NEC Labs America https://neclabs.wpengine.com/wp-content/uploads/2022/08/NEC-Labs-Blue-Logo-Square-300x267.jpg NEC Labs America2023-10-25 00:00:002023-11-16 20:13:18Calibrate Graph Neural Networks under Out-of-Distribution Nodes via Deep Q-learning
Graph neural networks (GNNs) have achieved great success in dealing with graph-structured data that are prevalent in the real world. The core of graph neural networks is the message passing mechanism that aims to generate the embeddings of nodes by aggregating the neighboring node information. However, recent work suggests that GNNs also suffer the trustworthiness issues. Our empirical study shows that the calibration error of the in-distribution (ID) nodes would be exacerbated if a graph is mixed with out-of-distribution (OOD) nodes, and we assume that the noisy information from OOD nodes is the root for the worsened calibration error. Both previous study and our empirical study suggest that adjusting the weights of edges could be a promising way to reduce the adverse impact from the OOD nodes. However, how to precisely select the desired edges and modify the corresponding weights is not trivial, since the distribution of OOD nodes is unknown to us. To tackle this problem, we propose a Graph Edge Re-weighting via Deep Q-learning (GERDQ) framework to calibrate the graph neural networks. Our framework aims to explore the potential influence of the change of the edge weights on target ID nodes by sampling and traversing the edges in the graph, and we formulate this process as a Markov Decision Process (MDP). Many existing GNNs could be seamlessly incorporated into our framework. Experimental results show that when wrapped with our method, the existing GNN models can yield lower calibration error under OOD nodes as well as comparable accuracy compared to the original ones and other strong baselines. The source code is available at:https://github.com/DamoSWL/Calibration-GNN-OOD.