<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://ovgu-nit.github.io/orakel/feed.xml" rel="self" type="application/atom+xml" /><link href="https://ovgu-nit.github.io/orakel/" rel="alternate" type="text/html" /><updated>2026-04-07T12:09:45+00:00</updated><id>https://ovgu-nit.github.io/orakel/feed.xml</id><title type="html">ORAKEL</title><author><name>NIT OVGU</name></author><entry><title type="html">Deep Learning-Based Gaze Estimation - A Review</title><link href="https://ovgu-nit.github.io/orakel/blog/2026/mdpi/" rel="alternate" type="text/html" title="Deep Learning-Based Gaze Estimation - A Review" /><published>2026-03-25T00:00:00+00:00</published><updated>2026-03-25T00:00:00+00:00</updated><id>https://ovgu-nit.github.io/orakel/blog/2026/MDPI</id><content type="html" xml:base="https://ovgu-nit.github.io/orakel/blog/2026/mdpi/"><![CDATA[<h2 id="deep-learning-based-gaze-estimation-a-review">Deep Learning-Based Gaze Estimation: A Review</h2>

<p>Gaze estimation, a critical facet of understanding user intent and enhancing human–computer interaction, has seen substantial advancements with the integration of deep learning technologies. Despite the progress, the application of deep learning in gaze estimation presents unique challenges, notably in the adaptation and optimization of these models for precise gaze tracking. This paper conducts a thorough review of recent developments in deep learning-based gaze estimation, with a particular focus on the evolution from traditional methods to sophisticated appearance-based techniques. We examine the key components of successful gaze estimation systems, including input feature processing, neural network architectures, and the importance of data preprocessing in achieving high accuracy. Our analysis extends to a comprehensive comparison of existing methods, shedding light on their effectiveness and limitations within various implementation contexts. Through this systematic review, we aim to consolidate existing knowledge in the field, identify gaps in current research, and suggest directions for future investigation. By providing a clear overview of the state-of-the-art in gaze estimation and discussing ongoing challenges and potential solutions, our work seeks to inspire further innovation and progress in developing more accurate and efficient gaze estimation systems.
<img src="/orakel/assets/theme/images/robotics-15-00069-g003.png" alt="" /></p>

<h2 id="fulltext-access">Fulltext Access</h2>
<p><a href="https://www.mdpi.com/2218-6581/15/4/69">https://www.mdpi.com/2218-6581/15/4/69</a></p>

<h2 id="citing">Citing</h2>
<div class="language-bibtex highlighter-rouge"><div class="highlight"><pre class="highlight"><code><span class="nc">@Article</span><span class="p">{</span><span class="nl">Abdelrahman2026</span><span class="p">,</span>

  <span class="na">author</span><span class="p">=</span><span class="s">{Abdelrahman, Ahmed and Al-Tawil, Basheer and Al-Hamadi, Ayoub}</span><span class="p">,</span>

  <span class="na">journal</span><span class="p">=</span><span class="s">{Robotics}</span><span class="p">,</span> 

  <span class="na">title</span><span class="p">=</span><span class="s">{Deep Learning-Based Gaze Estimation: A Review}</span><span class="p">,</span> 

  <span class="na">year</span><span class="p">=</span><span class="s">{2026}</span><span class="p">,</span>

  <span class="na">doi</span><span class="p">=</span><span class="s">{10.3390/robotics15040069}</span><span class="p">}</span>

</code></pre></div></div>]]></content><author><name>NIT OVGU</name></author><category term="Publication" /><category term="Journal" /><category term="MDPI" /><summary type="html"><![CDATA[Deep Learning-Based Gaze Estimation: A Review]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://ovgu-nit.github.io/orakel/assets/theme/images/robotics-15-00069-g003.png" /><media:content medium="image" url="https://ovgu-nit.github.io/orakel/assets/theme/images/robotics-15-00069-g003.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Die ORAKEL-Studie Magdeburg - Rezidivfrüherkennung bei Depression durch KI-gestützte Audio-/Videoanalyse</title><link href="https://ovgu-nit.github.io/orakel/blog/2026/aerzteblatt/" rel="alternate" type="text/html" title="Die ORAKEL-Studie Magdeburg - Rezidivfrüherkennung bei Depression durch KI-gestützte Audio-/Videoanalyse" /><published>2026-03-01T00:00:00+00:00</published><updated>2026-03-01T00:00:00+00:00</updated><id>https://ovgu-nit.github.io/orakel/blog/2026/Aerzteblatt</id><content type="html" xml:base="https://ovgu-nit.github.io/orakel/blog/2026/aerzteblatt/"><![CDATA[<h2 id="die-orakel-studie-magdeburg---rezidivfrüherkennung-bei-depression-durch-ki-gestützte-audio-videoanalyse">Die ORAKEL-Studie Magdeburg - Rezidivfrüherkennung bei Depression durch KI-gestützte Audio-/Videoanalyse</h2>

<p>Etwa 50 % der Patienten mit Depression erleben trotz Behandlung einen Rückfall. Die ORAKEL-Studie in Magdeburg testet, ob multimodale Audio-/Videoanalysen die Früherkennung dieser Rückfälle verbessern. Über einen Zeitraum von 48 Wochen werden sechs ärztlich-psychologische Verlaufsuntersuchungen durchgeführt, bei denen parallel Sprach-, Mimik- und Vitalparameter erfasst werden. Künstliche Intelli- genz (KI) Modelle sollen entwickelt werden, die Veränderungen erkennen, die sich leicht dem ärztlichen Blick entziehen – zur intelligenten Unterstützung, nicht zum Ersatz der klinischen Beurteilung.
<img src="/vitalsigns/assets/theme/images/fb-26-3-af3d4147.webp" alt="" /></p>

<h2 id="fulltext-access">Fulltext Access</h2>
<p><a href="https://www.aerzteblatt-sachsen-anhalt.de/component/content/article/die-orakel-studie-magdeburg-fb-2026-03">https://www.aerzteblatt-sachsen-anhalt.de/component/content/article/die-orakel-studie-magdeburg-fb-2026-03</a></p>]]></content><author><name>NIT OVGU</name></author><category term="Publication" /><category term="Journal" /><category term="Ärzteblatt" /><summary type="html"><![CDATA[Die ORAKEL-Studie Magdeburg - Rezidivfrüherkennung bei Depression durch KI-gestützte Audio-/Videoanalyse]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://ovgu-nit.github.io/orakel/assets/theme/images/fb-26-3-af3d4147.webp" /><media:content medium="image" url="https://ovgu-nit.github.io/orakel/assets/theme/images/fb-26-3-af3d4147.webp" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">STCM-Mamba - Multimodal Spatio-Temporal Cross-Modal Mamba for Depression Detection</title><link href="https://ovgu-nit.github.io/orakel/blog/2025/ieee/" rel="alternate" type="text/html" title="STCM-Mamba - Multimodal Spatio-Temporal Cross-Modal Mamba for Depression Detection" /><published>2025-10-31T00:00:00+00:00</published><updated>2025-10-31T00:00:00+00:00</updated><id>https://ovgu-nit.github.io/orakel/blog/2025/IEEE</id><content type="html" xml:base="https://ovgu-nit.github.io/orakel/blog/2025/ieee/"><![CDATA[<h2 id="stcm-mamba-multimodal-spatio-temporal-cross-modal-mamba-for-depression-detection">STCM-Mamba: Multimodal Spatio-Temporal Cross-Modal Mamba for Depression Detection</h2>

<p>Depression is a prevalent mental disorder with severe physiological symptoms and high diagnosis costs, however, the development of efficient and accurate depression detection systems remains challenging. While deep learning methods leveraging multimodal data have shown promise, existing approaches suffer from two critical limitations. One is the lack of effective spatiotemporal feature integration, and the other is the inability to balance effective long-sequence modeling and computational complexity. To address these challenges, we propose STCM-Mamba, a novel Spatio-Temporal Cross-Modal Mamba (STCM-Mamba) framework for efficient and accurate depression detection. The STCM-Mamba comprises three modules: a Spatio-Temporal Mamba Module (STMM), a Cross-Modal Mamba Module (CMMM), and a Depression Classification Module (DCM). The STMM consists of a Temporal Mamba Block (TMB) and a Spatial Mamba Block (SMB) for capturing spatiotemporal information for each modality, while the CMMM enhances intermodal and intramodal representation learning. Experiments on two multimodal depression datasets demonstrate that STMM and CMMM significantly contribute to performance improvements, and the STCM-Mamba outperforms state-of-the-art methods.
<img src="/orakel/assets/theme/images/access-gagraphic-3627778.jpg" alt="" /></p>

<h2 id="fulltext-access">Fulltext Access</h2>
<p><a href="https://ieeexplore.ieee.org/document/11223210">https://ieeexplore.ieee.org/document/11223210</a></p>

<h2 id="citing">Citing</h2>
<div class="language-bibtex highlighter-rouge"><div class="highlight"><pre class="highlight"><code><span class="nc">@Article</span><span class="p">{</span><span class="nl">Zhou2025</span><span class="p">,</span>

  <span class="na">author</span><span class="p">=</span><span class="s">{Zhou, Bowen and Fiedler, Marc-André and Al-Hamadi, Ayoub}</span><span class="p">,</span>

  <span class="na">journal</span><span class="p">=</span><span class="s">{IEEE Access}</span><span class="p">,</span> 

  <span class="na">title</span><span class="p">=</span><span class="s">{STCM-Mamba: Multimodal Spatio-Temporal Cross-Modal Mamba for Depression Detection}</span><span class="p">,</span> 

  <span class="na">year</span><span class="p">=</span><span class="s">{2025}</span><span class="p">,</span>

  <span class="na">doi</span><span class="p">=</span><span class="s">{10.1109/ACCESS.2025.3627778}</span><span class="p">}</span>

</code></pre></div></div>]]></content><author><name>NIT OVGU</name></author><category term="Publication" /><category term="Journal" /><category term="IEEE" /><summary type="html"><![CDATA[STCM-Mamba: Multimodal Spatio-Temporal Cross-Modal Mamba for Depression Detection]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://ovgu-nit.github.io/orakel/assets/theme/images/access-gagraphic-3627778.jpg" /><media:content medium="image" url="https://ovgu-nit.github.io/orakel/assets/theme/images/access-gagraphic-3627778.jpg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Cross-Modal Fusion Mamba for Multimodal Depression Detection</title><link href="https://ovgu-nit.github.io/orakel/blog/2025/ispa/" rel="alternate" type="text/html" title="Cross-Modal Fusion Mamba for Multimodal Depression Detection" /><published>2025-10-29T00:00:00+00:00</published><updated>2025-10-29T00:00:00+00:00</updated><id>https://ovgu-nit.github.io/orakel/blog/2025/ISPA</id><content type="html" xml:base="https://ovgu-nit.github.io/orakel/blog/2025/ispa/"><![CDATA[<h2 id="cross-modal-fusion-mamba-for-multimodal-depression-detection">Cross-Modal Fusion Mamba for Multimodal Depression Detection</h2>

<p>Depression detection using multimodal signals has garnered growing attention due to its potential for early warning. However, existing approaches often rely on limited visual features or computationally intensive fusion mechanisms. In this study, we present a novel framework based on the Mamba structure to address these challenges. Our method fuses audio features with enriched visual representations by combining facial landmarks and action units (AUs), enhancing the expressiveness of visual cues. To capture intramodal information, we propose the Audio Mamba Encoder (AME) for audio modality, and the Vision CrossMamba (VCM) module for visual feature fusion. Furthermore, the Audio-Vision CrossMamba (AVCM) module is designed for intermodal interactions. Experimental results demonstrate superior performance over several baselines, highlighting the effectiveness of the proposed framework in detecting depression from multimodal data.
<img src="/orakel/assets/theme/images/11259458-fig-1-source-large.gif" alt="" /></p>

<h2 id="fulltext-access">Fulltext Access</h2>
<p><a href="https://ieeexplore.ieee.org/document/11259458">https://ieeexplore.ieee.org/document/11259458</a></p>

<h2 id="citing">Citing</h2>
<div class="language-bibtex highlighter-rouge"><div class="highlight"><pre class="highlight"><code><span class="nc">@INPROCEEDINGS</span><span class="p">{</span><span class="nl">Zhou2025</span><span class="p">,</span>

  <span class="na">author</span><span class="p">=</span><span class="s">{Zhou, Bowen and Fiedler, Marc-André and Al-Hamadi, Ayoub}</span><span class="p">,</span>

  <span class="na">booktitle</span><span class="p">=</span><span class="s">{2025 14th International Symposium on Image and Signal Processing and Analysis (ISPA)}</span><span class="p">,</span> 

  <span class="na">title</span><span class="p">=</span><span class="s">{Cross-Modal Fusion Mamba for Multimodal Depression Detection}</span><span class="p">,</span> 

  <span class="na">year</span><span class="p">=</span><span class="s">{2025}</span><span class="p">,</span>

  <span class="na">doi</span><span class="p">=</span><span class="s">{10.1109/ISPA66905.2025.11259458}</span><span class="p">}</span>

</code></pre></div></div>]]></content><author><name>NIT OVGU</name></author><category term="Publication" /><category term="Conference" /><category term="ISPA" /><summary type="html"><![CDATA[Cross-Modal Fusion Mamba for Multimodal Depression Detection]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://ovgu-nit.github.io/orakel/assets/theme/images/11259458-fig-1-source-large.gif" /><media:content medium="image" url="https://ovgu-nit.github.io/orakel/assets/theme/images/11259458-fig-1-source-large.gif" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Multi-Modal AI-Based Pain Detection in Intermediate Care Patients in the Postoperative Phase</title><link href="https://ovgu-nit.github.io/orakel/blog/2025/smc/" rel="alternate" type="text/html" title="Multi-Modal AI-Based Pain Detection in Intermediate Care Patients in the Postoperative Phase" /><published>2025-10-08T00:00:00+00:00</published><updated>2025-10-08T00:00:00+00:00</updated><id>https://ovgu-nit.github.io/orakel/blog/2025/SMC</id><content type="html" xml:base="https://ovgu-nit.github.io/orakel/blog/2025/smc/"><![CDATA[<h2 id="multi-modal-ai-based-pain-detection-in-intermediate-care-patients-in-the-postoperative-phase">Multi-Modal AI-Based Pain Detection in Intermediate Care Patients in the Postoperative Phase</h2>

<p>“Multi-modal AI-Based Pain Detection in Intermediate Care Patients in the Postoperative Phase” is an interdisciplinary research work that operates in the domain of automated pain detection. It aims to improve previous work, based on pain databases like BioVid and UNBC shoulder pain, as well as AI-based approaches using computer vision and signal processing to analyze available modalities. Thus, we present our basic research idea on how to improve automatic pain detection in three major steps. The first step focuses on collecting pain data from postoperative patients in intermediate care stations (IMC). In addition, patients who are not fully oriented should be included in a separate data collection as a second focus group. Then, improvements on the state-of-the-art models should not only advance general pain detection, but also help bridge the gap to the real-world setting of the IMC data. Improvements include transferability analysis, feature selection evaluation, and balancing of data distribution to deliver better classification performance. In a last step, we aim to test, verify and evaluate the classification performance on the IMC data with the support of medical practitioners.
<img src="/orakel/assets/theme/images/niena1-p4-niena-large.gif" alt="" /></p>

<h2 id="fulltext-access">Fulltext Access</h2>
<p><a href="https://ieeexplore.ieee.org/document/11343734">https://ieeexplore.ieee.org/document/11343734</a></p>

<h2 id="citing">Citing</h2>
<div class="language-bibtex highlighter-rouge"><div class="highlight"><pre class="highlight"><code><span class="nc">@INPROCEEDINGS</span><span class="p">{</span><span class="nl">Nienaber2025</span><span class="p">,</span>

  <span class="na">author</span><span class="p">=</span><span class="s">{Nienaber, Sören and Wang, Huibin and Hempel, Thorsten and Walter, Steffen and Barth, Eberhard and Al-Hamadi, Ayoub}</span><span class="p">,</span>

  <span class="na">booktitle</span><span class="p">=</span><span class="s">{2025 IEEE International Conference on Systems, Man, and Cybernetics (SMC)}</span><span class="p">,</span> 

  <span class="na">title</span><span class="p">=</span><span class="s">{Multi-Modal AI-Based Pain Detection in Intermediate Care Patients in the Postoperative Phase}</span><span class="p">,</span> 

  <span class="na">year</span><span class="p">=</span><span class="s">{2025}</span><span class="p">,</span>

  <span class="na">doi</span><span class="p">=</span><span class="s">{10.1109/SMC58881.2025.11343734}</span><span class="p">}</span>

</code></pre></div></div>]]></content><author><name>NIT OVGU</name></author><category term="Publication" /><category term="Conference" /><category term="SMC" /><summary type="html"><![CDATA[Multi-Modal AI-Based Pain Detection in Intermediate Care Patients in the Postoperative Phase]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://ovgu-nit.github.io/orakel/assets/theme/images/niena1-p4-niena-large.gif" /><media:content medium="image" url="https://ovgu-nit.github.io/orakel/assets/theme/images/niena1-p4-niena-large.gif" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Patient Data Recording Start</title><link href="https://ovgu-nit.github.io/orakel/blog/2025/recording/" rel="alternate" type="text/html" title="Patient Data Recording Start" /><published>2025-04-01T00:00:00+00:00</published><updated>2025-04-01T00:00:00+00:00</updated><id>https://ovgu-nit.github.io/orakel/blog/2025/recording</id><content type="html" xml:base="https://ovgu-nit.github.io/orakel/blog/2025/recording/"><![CDATA[<h2 id="data-collection-at-university-hospital-for-psychiatry-and-psychotherapy-begins">Data Collection at University Hospital for Psychiatry and Psychotherapy Begins</h2>

<p>We are pleased to announce that the data collection phase has officially begun at our partner institution, the University Hospital for Psychiatry and Psychotherapy. The study will involve 120 patients, with each patient attending six scheduled sessions.</p>

<p>Each session consists of:</p>

<ul>
  <li>
    <p>A medical-psychiatric interview</p>
  </li>
  <li>
    <p>A psychological interview</p>
  </li>
  <li>
    <p>Standardized clinical ratings</p>
  </li>
</ul>

<p>The data collection takes place in a specially equipped room, featuring state-of-the-art hardware for recording purposes. To ensure comprehensive data capture, three cameras are used during each session:</p>

<ul>
  <li>
    <p>One focusing on the patient</p>
  </li>
  <li>
    <p>One on the doctor or psychologist</p>
  </li>
  <li>
    <p>One monitoring the patient’s gait when entering and leaving the room</p>
  </li>
</ul>

<p>The data collected will enable the extraction and analysis of various features for detailed behavioral and condition assessment. This includes:</p>

<ul>
  <li>
    <p>Facial expressions</p>
  </li>
  <li>
    <p>Head and body posture</p>
  </li>
  <li>
    <p>Emotions</p>
  </li>
  <li>
    <p>Gait analysis</p>
  </li>
  <li>
    <p>Vital signs</p>
  </li>
  <li>
    <p>Speech patterns</p>
  </li>
</ul>

<p>This initiative marks a significant step forward in our collaborative efforts to better understand psychiatric conditions and improve patient care through innovative technology and research.</p>

<p><img src="/orakel/assets/theme/images/layout_recording.png" alt="" /></p>]]></content><author><name>NIT OVGU</name></author><category term="News" /><category term="News" /><summary type="html"><![CDATA[Data Collection at University Hospital for Psychiatry and Psychotherapy Begins]]></summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://ovgu-nit.github.io/orakel/assets/theme/images/layout_recording_cropped.png" /><media:content medium="image" url="https://ovgu-nit.github.io/orakel/assets/theme/images/layout_recording_cropped.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry></feed>