Skip to content

Use Cases & Applications

ChronosVector’s temporal-native design serves a range of domains where understanding how embeddings evolve over time is as important as knowing where they are right now. This page describes four detailed use cases with domain context, the CVX features each relies on, and example query flows.


Early detection of psychological disorders (depression, eating disorders, self-harm risk) from social media is a growing area of clinical NLP research. The core challenge: individual posts are noisy and unreliable indicators, but the trajectory of a user’s language over weeks or months reveals meaningful patterns — gradual vocabulary shifts, increasing negativity, social withdrawal reflected in changing topics.

Traditional approaches embed each post independently and compute aggregate statistics (mean sentiment, keyword frequency). This discards the temporal structure that is often the strongest signal.

User → Posts over time → Embed each post → Trajectory in embedding space
t1: "Had a great day..." → v(t1)
t2: "Feeling tired again..." → v(t2)
t3: "Nothing matters..." → v(t3)
...

Each user becomes an entity in CVX with a trajectory of post embeddings indexed by timestamp. The trajectory captures the evolution of their language patterns over time.

Velocity and acceleration quantify the rate and direction of linguistic change. A user whose embedding velocity suddenly increases may be entering a crisis period. The drift significance test distinguishes genuine directional change from natural linguistic variation.

Change point detection (PELT for retrospective analysis, BOCPD for real-time monitoring) identifies moments when a user’s language patterns shift abruptly — potential onset or escalation events.

Drift attribution explains what changed in the embedding space, mapping high-dimensional shifts to interpretable dimensions that clinicians can act on.

Prefix features enable early detection: CVX can compute trajectory features (velocity, volatility, Hurst exponent) on the prefix of a trajectory — the data available so far — and compare against known risk profiles without waiting for complete histories.

Path signatures classify trajectory shapes. A signature captures whether a user’s language is trending, cycling, or stable, regardless of the specific content. Signature-based similarity search finds users with similar evolution patterns, enabling cohort analysis.

CVX’s explain layer translates raw analytics into clinician-friendly narratives. Instead of a 768-dimensional velocity vector, a clinician sees: “User’s language shifted significantly (p < 0.01) toward social withdrawal topics between March 15 and April 2, with a change point detected on March 22.”

This use case is grounded in the LabChain case study for temporal word embeddings applied to early detection of psychological disorders on social media (Couto et al., 2025).


Quantitative researchers face several temporal embedding challenges:

  • Market regime matching: Find historical periods whose market structure resembles the current state. Traditional approaches compare scalar indicators (VIX, yield spreads); embedding-based approaches capture the full high-dimensional structure of the market.
  • Factor decay detection: Quantitative factors (momentum, value, quality) lose predictive power over time. Detecting this decay early — before it erodes returns — requires monitoring the evolution of factor embeddings.
  • Path-dependent pricing: Derivatives whose payoff depends on the path of an underlying (Asian options, lookbacks, barrier options) benefit from path signature representations.

Temporal kNN with regime detection finds historical periods similar to the current market state. A quant researcher embeds market data (returns, correlations, macro indicators) into a high-dimensional space and queries CVX for the nearest historical periods, weighted by both semantic similarity and temporal proximity.

Mean reversion tests and GARCH volatility characterize factor dynamics. A factor embedding that tests as mean-reverting with a short half-life is likely to recover from a drawdown. One that tests as a random walk may be permanently decaying. GARCH persistence tells you whether current volatility is a temporary spike or a regime shift.

Path signatures provide fixed-dimensional descriptors of market trajectories that are invariant to reparametrization — a property that aligns with the intuition that the shape of a market move matters more than its exact timing.

Hurst exponent distinguishes trending markets (H>0.5H > 0.5) from mean-reverting ones (H<0.5H < 0.5), informing whether momentum or contrarian strategies are appropriate.

  1. Embed current market state (e.g., cross-sectional factor returns) as a vector vnowv_{\text{now}}
  2. Query CVX: temporal_knn(v_now, k=10, time_weight=0.3) — find similar historical states
  3. For each matched period, retrieve the subsequent trajectory: trajectory(entity, t_match, t_match + 60d)
  4. Compute signatures on those forward trajectories to identify the dominant post-match pattern
  5. Use the pattern distribution to inform position sizing and risk management

CVX follows the domain-agnostic core principle (ADR-015): the system does not contain finance-specific logic. Mean reversion tests, GARCH models, and path signatures are general-purpose tools for trajectory analysis. The finance interpretation arises from the data and queries, not from CVX’s code. This same infrastructure serves NLP, clinical, and MLOps use cases without modification.

This philosophy aligns with Lopez de Prado’s information-driven sampling approach: let the data’s structure drive the analysis, rather than imposing domain-specific assumptions.


Production ML systems suffer from embedding drift — the gradual divergence between the embedding distribution a model was trained on and the distribution it encounters in production. This drift degrades model performance silently, often detected only when downstream metrics (click-through rate, conversion, accuracy) have already deteriorated.

Current monitoring tools track scalar metrics (accuracy, latency, feature distributions) but lack the ability to analyze high-dimensional embedding trajectories directly.

Embedding drift tracking across model versions. Each model version produces embeddings that CVX stores with timestamps and model version metadata. CVX’s cohort divergence analytics quantify how embeddings from version NN and version N+1N+1 diverge over time.

Declarative drift monitors trigger alerts when embedding drift exceeds thresholds. CVX computes velocity, volatility, and change points in real time via BOCPD, generating DriftEvent notifications through the gRPC WatchDrift stream.

Model version alignment compares embedding spaces across model versions. When a model is retrained, its embedding space may rotate or shift. CVX can compute the transformation between versions and track whether the alignment holds over time.

Provenance tracking records which model, pipeline, and data version produced each embedding. When drift is detected, engineers can trace back to the exact model version and training data that generated the drifting embeddings.

  1. Ingest: Production embeddings flow into CVX via the gRPC IngestStream, tagged with model version and timestamp
  2. Monitor: A declarative monitor watches for velocity exceeding a threshold: “Alert if embedding drift rate > 0.05 for any entity cohort over a 24h window”
  3. Diagnose: When alerted, the engineer queries CVX for drift attribution: which dimensions shifted, when the change point occurred, and whether the shift is statistically significant
  4. Compare: Side-by-side trajectory comparison between the current model and the previous stable version identifies whether the drift is in the model or the data
  5. Act: If the drift is data-driven (new topics, distribution shift), retrain. If model-driven (training instability), roll back
  • POST /v1/ingest with model version metadata
  • GET /v1/entities/{id}/velocity for real-time drift rate
  • GET /v1/entities/{id}/changepoints for shift detection
  • gRPC WatchDrift for streaming alerts
  • Cohort divergence for cross-version comparison
  • Provenance queries for root cause analysis

Diachronic semantics — the study of how word meanings change over time — is a core area of computational linguistics. Researchers study questions like: How did “cloud” shift from weather to computing? When did “transformer” acquire its ML sense? Do high-frequency words change faster than low-frequency ones?

Traditional approaches train separate embedding models per time period and align them post-hoc (Hamilton et al., 2016; Yao et al., 2018). This introduces alignment artifacts and loses fine-grained temporal resolution.

Native temporal storage eliminates the alignment problem. CVX stores embeddings with continuous timestamps, so researchers can query any time point without binning into discrete periods.

Temporal analogy queries answer questions of the form: “What in 2018 played the role that X plays in 2024?” This is a temporal generalization of the classic analogy task. CVX computes this by finding entities whose position at t1t_1 matches the query entity’s position at t2t_2, after accounting for the global shift of the embedding space.

Trajectory analysis over a concept’s full history reveals patterns invisible to snapshot comparisons:

  • Velocity quantifies the rate of semantic change over time, testing Hamilton et al.’s laws of conformity and innovation
  • Change points identify when a word acquired a new meaning (e.g., “streaming” transitioning from video to music)
  • Mean reversion tests determine whether a semantic shift is permanent or temporary
  • Hurst exponent reveals whether semantic change is persistent (trending) or anti-persistent (self-correcting)

Cohort divergence tracks how groups of related concepts evolve together or apart. Do all programming language names drift together? Does the “AI” cohort (neural network, deep learning, transformer, attention) converge or diverge over time?

  1. Load temporal embeddings for a corpus (e.g., Wikipedia monthly snapshots 2018-2025)
  2. Query a concept’s trajectory: GET /v1/entities/transformer/trajectory?t1=2017-01&t2=2025-01
  3. Detect change points: GET /v1/entities/transformer/changepoints — identifies when the ML sense emerged
  4. Run temporal analogy: “What in 2015 occupied the embedding position that ‘transformer’ occupies in 2024?” — returns “LSTM”, “RNN”
  5. Compare cohort evolution: track how “transformer”, “attention”, “BERT”, “GPT” move as a group
  6. Compute path signatures for trajectory classification: which concepts follow a “gradual broadening” pattern vs. “abrupt shift” pattern?
  • Temporal kNN with timestamp weighting
  • Trajectory retrieval with continuous timestamps
  • Change point detection (PELT for offline, BOCPD for streaming corpora)
  • Temporal analogy queries
  • Cohort divergence analytics
  • Path signatures for trajectory classification
  • Stochastic process characterization (drift, volatility, Hurst, stationarity)