Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ETP: Learning Transferable ECG Representations via ECG-Text Pre-training (2309.07145v1)

Published 6 Sep 2023 in eess.SP, cs.AI, and cs.LG

Abstract: In the domain of cardiovascular healthcare, the Electrocardiogram (ECG) serves as a critical, non-invasive diagnostic tool. Although recent strides in self-supervised learning (SSL) have been promising for ECG representation learning, these techniques often require annotated samples and struggle with classes not present in the fine-tuning stages. To address these limitations, we introduce ECG-Text Pre-training (ETP), an innovative framework designed to learn cross-modal representations that link ECG signals with textual reports. For the first time, this framework leverages the zero-shot classification task in the ECG domain. ETP employs an ECG encoder along with a pre-trained LLM to align ECG signals with their corresponding textual reports. The proposed framework excels in both linear evaluation and zero-shot classification tasks, as demonstrated on the PTB-XL and CPSC2018 datasets, showcasing its ability for robust and generalizable cross-modal ECG feature learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Che Liu (59 papers)
  2. Zhongwei Wan (39 papers)
  3. Sibo Cheng (36 papers)
  4. Mi Zhang (85 papers)
  5. Rossella Arcucci (50 papers)
Citations (26)

Summary

We haven't generated a summary for this paper yet.