Loading Now

Summary of Preventing Conflicting Gradients in Neural Marked Temporal Point Processes, by Tanguy Bosser and Souhaib Ben Taieb


Preventing Conflicting Gradients in Neural Marked Temporal Point Processes

by Tanguy Bosser, Souhaib Ben Taieb

First submitted to arxiv on: 11 Dec 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: None

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
The paper proposes a new approach to learning Neural Marked Temporal Point Processes (MTPP) models that capture complex temporal inter-dependencies between labeled events. The authors frame MTPP learning as a two-task problem, where both tasks share trainable parameters optimized jointly. This leads to conflicting gradients during training, which can be detrimental to task performance. To overcome this issue, the authors introduce novel parametrizations for neural MTPP models that allow separate modeling and training of each task. The framework is evaluated on multiple real-world event sequence datasets, demonstrating improved performance compared to original model formulations.
Low GrooveSquid.com (original content) Low Difficulty Summary
The paper teaches us about a new way to understand events happening in time. It’s like trying to predict when and what will happen next. The old way was complicated because it tried to do two things at once: predict the timing of events and what type of event it would be. This caused problems during training, making it hard for the model to learn correctly. To fix this, the authors created a new way to train these models that lets them focus on each task separately. They tested their method on real-world data and showed that it works better than the old way.

Keywords

» Artificial intelligence