Papers
Topics
Authors
Recent
2000 character limit reached

Double-Tag Technique in Watermarking & Classification

Updated 11 January 2026
  • Double-Tag Technique is a method that uses two independent tagging phases in both digital watermarking and data-driven classifier training to enhance robustness and detectability.
  • In watermarking, the approach embeds watermarks via DCT and SVD, achieving high PSNR (≈52.5 dB) even after noise attacks and ensuring reliable extraction.
  • For classifier training, the Tag N’ Train method iteratively refines weak taggers into strong classifiers, improving metrics such as AUC from 0.65 to 0.82 in practical experiments.

The double-tag technique refers to two distinct but structurally parallel methodologies for improving robustness and performance in two research domains: (1) digital watermarking via DCT and SVD dual-embedding, and (2) data-driven classifier training via the Tag N’ Train (TNT) bootstrapping scheme. Both approaches utilize two independent “tagging” or embedding phases—one to enhance imperceptibility and resilience, the other to leverage complementary information for inference or classification. This entry synthesizes the formal models, mathematical foundations, algorithmic steps, and experimental outcomes of these methods as established by the seminal works "A DCT And SVD based Watermarking Technique To Identify Tag" (Ji et al., 2015) and "Tag N' Train: A Technique to Train Improved Classifiers on Unlabeled Data" (Amram et al., 2020).

1. Double-Embedding for Digital Watermarking: DCT+SVD Scheme

The double-tag watermarking approach for images involves consecutive embedding of watermark information in both the frequency domain and the singular value domain. The process is comprised of two main stages:

  • DCT-Based Embedding: The cover image ff of size M×NM \times N is segmented into 8×88 \times 8 blocks. For each block fn(x,y)f_n(x,y), the 2D DCT is applied:

Xn(u,v)=x=07y=07fn(x,y)cos ⁣((2x+1)uπ16)cos ⁣((2y+1)vπ16)X_n(u,v) = \sum_{x=0}^7 \sum_{y=0}^7 f_n(x,y)\,\cos\!\left(\frac{(2x+1)u\pi}{16}\right) \cos\!\left(\frac{(2y+1)v\pi}{16}\right)

The watermark is embedded by adding scaled watermark bits wiw_i to selected mid-band diagonal coefficients:

Xn(ui,ui)=Xn(ui,ui)+αwi,α=0.05X_n'(u_i,u_i) = X_n(u_i,u_i) + \alpha w_i,\qquad \alpha=0.05

The inverse DCT reconstructs the intermediate watermarked image ff'.

  • SVD-Based Embedding: The intermediate image ff' is again blockwise decomposed, and each block AnA_n undergoes SVD:

An=UnΣnVnTA_n = U_n \Sigma_n V_n^T

The same or a different kk-bit watermark is embedded into the singular values:

Σn=Σn+αWn\Sigma_n' = \Sigma_n + \alpha W_n

WnW_n is a diagonal matrix encoding the watermark bits. Each block is recomposed to obtain the final doubly watermarked image AA''.

This architecture ensures that the watermark information resides simultaneously in two distinct transform domains (DCT and SVD), strengthening resistance to removal or distortion in either domain alone.

2. Extraction, Performance Criteria, and Robustness

Extraction operates in reverse:

  • SVD-Phase Extraction: Each watermarked image block AnA_n'' undergoes SVD:

An=U~nΣnV~nTA_n'' = \tilde{U}_n \Sigma_n'' \tilde{V}_n^T

The embedded bits are estimated by:

wi=σniσniαw_i = \frac{\sigma_n''^i - \sigma_n^i}{\alpha}

where σni\sigma_n^i are the singular values from the intermediate image ff'.

  • DCT-Phase Extraction: After SVD extraction and optional reconstruction of ff', DCT is again applied blockwise to extract watermark bits from mid-band diagonal coefficients as:

wi=Xn(ui,ui)Xn(ui,ui)αw_i = \frac{X_n''(u_i,u_i) - X_n(u_i,u_i)}{\alpha}

Robustness is quantified using Peak Signal-to-Noise Ratio (PSNR),

PSNR=10log10(MAXI2MSE)\text{PSNR} = 10 \log_{10} \left( \frac{\mathrm{MAX}_I^2}{\mathrm{MSE}} \right)

where MAXI=255\mathrm{MAX}_I=255 for 8-bit images and MSE is mean squared error. For α=0.05\alpha=0.05, typical PSNR is 52.49 dB (imperceptible distortion). Even after attacks (Gaussian, salt-and-pepper noise), PSNR remains above 35 dB, and watermark recoverability is preserved. The threshold PSNRwm>T_{\mathrm{wm}}>T (e.g., T=30T=30 dB) is adopted for authenticity verification (Ji et al., 2015).

3. Double-Tag Bootstrapping for Data-Driven Classifier Training

The TNT (Tag N’ Train) methodology applies a two-phase tagging protocol to exploit structure in unlabeled datasets, particularly events with two correlated sub-objects, such as dijet events in collider physics. The core procedure is:

  • Phase 1: Weak Tagging. A weak classifier f1:xi[0,1]f_1: x_i \rightarrow [0,1] is deployed on object ii in every event. Thresholds τs>τb\tau_s > \tau_b partition the dataset into signal-rich (S1S_1) and background-rich (B1B_1) samples:

S1={kf1(xik)τs},B1={kf1(xik)τb}S_1 = \{k \mid f_1(x_i^k) \geq \tau_s\}, \quad B_1 = \{k \mid f_1(x_i^k) \leq \tau_b\}

In collider settings, f1f_1 is often an autoencoder with reconstruction loss LAE(x)L_{AE}(x), and tags are defined by quantiles of LAEL_{AE}.

  • Phase 2: Strong Classifier Training. A classifier f2:xj[0,1]f_2: x_j \rightarrow [0,1] is trained to distinguish S1S_1 (pseudo-signal, y=1y'=1) from B1B_1 (pseudo-background, y=0y'=0) using a weighted cross-entropy loss:

L(f2)=[wSxjS1logf2(xj)+wBxjB1log(1f2(xj))]L(f_2) = -\left[w_S \sum_{x_j \in S_1} \log f_2(x_j) + w_B \sum_{x_j \in B_1} \log(1 - f_2(x_j))\right]

Sample weights wSw_S and wBw_B normalize for class imbalance.

Iterative refinement optionally alternates roles, using f2f_2 to retag object jj and retrain f1f_1, with multiple cycles (TT iterations). This bootstraps performance as the mixture in S1S_1, B1B_1 becomes increasingly pure (Amram et al., 2020).

4. Theoretical and Algorithmic Foundations

In the TNT method, the mixture model analysis ensures that, provided the weak tagger satisfies P(tagsSignal)P(tagsBackground)P(\text{tag}_s|\text{Signal}) \gg P(\text{tag}_s|\text{Background}) and vice versa, the distribution p(xjTagged=S)p(x_j|\text{Tagged}=S) is dominated by signal in xjx_j. Thus, f2f_2 asymptotically approaches the likelihood ratio optimal for distinguishing signal from background:

Training f2 to separate S1 from B1f2(xj)P(Signalxj)\text{Training } f_2\text{ to separate } S_1\text{ from } B_1 \Rightarrow f_2(x_j) \approx P(\text{Signal}|x_j)

In the watermarking context, double embedding leverages the orthogonality of DCT basis vectors and the stability of SVD singular values. Each embedding operation is algebraically orthogonal but not non-interfering, so the second embedding can induce perturbation to the first. This is managed by explicit storage or regeneration of original coefficients and careful setting of α\alpha.

5. Practical Implementation and Experimental Results

Experimental setups in both domains demonstrate the utility of double-tag schemes.

  • Watermarking (DCT+SVD): For an embedding strength α=0.05\alpha=0.05, watermarked images maintain PSNR ≈ 52.49 dB. After various noise attacks, PSNR degrades to 35.88–49.30 dB, yet both DCT and SVD watermarks remain detectable. The scheme resists attempts to remove a watermark by targeting a single domain, as redundancy across orthogonal representations ensures residual retrievability (Ji et al., 2015).
  • TNT Classifier Training: A worked example for LHC dijet resonance searches uses a convolutional autoencoder as f1f_1 and a CNN as f2f_2, trained on 200k unlabeled events/iteration. In the LHC-Olympics dijet benchmark (1% signal), AUC of f2f_2 is elevated from 0.65 (autoencoder) to 0.82, and statistical significance for discovery is raised from ≈3σ to ≫5σ. Performance can approach that of fully supervised training if the sub-object independence assumption is satisfied (Amram et al., 2020).
Domain Implementation Steps Typical Outcomes
DCT+SVD Watermarking Block DCT → diag tag → block SVD PSNR>50 dB, robust under noise
Tag N’ Train (TNT) Weak tag → strong classifier training AUC ∼0.8+, σ_discovery≫5σ

6. Advantages, Limitations, and Future Directions

Advantages

  • In watermarking, double-tag embedding in both DCT and SVD domains yields low visual distortion pre-attack, and high resilience against additive/impulse noise. This redundancy impedes successful attacks targeting only a single domain.
  • In data classification, TNT allows model-agnostic, simulation-independent classifier training directly on unlabeled data, well-suited for anomaly detection in two-object systems.

Limitations

  • In watermarking, the dual embedding increases computational overhead and can induce minor cumulative degradation of the first tag. Storage of either original DCT coefficients or singular values is necessary.
  • In TNT, performance depends critically on the weak tagger’s discriminative ability and the statistical independence of sub-object features on background-only events. The method assumes two or more sub-objects per event and lacks single-object generality.

Extensions

  • TNT may incorporate soft (score-based) labeling, alternative unsupervised taggers (e.g., normalizing flows), and can generalize to multi-tag settings with more than two sub-objects.
  • Adversarial decorrelation can enforce independence from kinematic variables, such as jet pTp_T in particle physics applications.

7. Comparative Summary

Both analyzed double-tag approaches establish that embedding, tagging, or training in two orthogonal domains or subspaces significantly enhances system robustness. In image watermarking, redundancy across frequency and singular-value subspaces fortifies watermark integrity. In unsupervised learning, bootstrapping from weak to strong taggers transforms ambiguous labels into increasingly pure classification signals. The trade-off universally observed is between increased algorithmic complexity and the gain in reliability and interpretability.

Key results and methodologies are detailed in "A DCT And SVD based Watermarking Technique To Identify Tag" (Ji et al., 2015) and "Tag N' Train: A Technique to Train Improved Classifiers on Unlabeled Data" (Amram et al., 2020).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Whiteboard

Topic to Video (Beta)

Follow Topic

Get notified by email when new papers are published related to Double-Tag Technique.