Skip to content

TypePlus is a lightweight deep learning model for free-text keystroke authentication, evaluated across multiple loss functions, including a Filtered Mean Triplet Loss. It achieves state-of-the-art performance with a 2.86% Equal Error Rate (EER) on the Aalto University Keystroke Dataset, demonstrating its effectiveness in keystroke biometrics.

Notifications You must be signed in to change notification settings

AhmedYasserrr/TypePlus

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 

Repository files navigation

TypePlus: A Deep Learning Architecture for Keystroke Authentication with Loss Function Evaluation

This repository provides the implementation and dataset for the paper "TypePlus: A Deep Learning Architecture for Keystroke Authentication with Loss Function Evaluation", submitted to the 2nd International Conference on Intelligent Systems, Blockchain, and Communication Technologies (ISBCom) 2025.

Table of Contents

  1. Dataset
  2. Data Preprocessing
  3. Model Architecture
  4. Results
  5. Project Setup
  6. Running the Experiment

TypePlus is a lightweight deep learning model for free-text keystroke authentication, evaluated across multiple loss functions, including a Filtered Mean Triplet Loss. It achieves state-of-the-art performance with a 2.86% Equal Error Rate (EER) on the Aalto University Keystroke Dataset, demonstrating its effectiveness in keystroke biometrics.

1. Dataset

TypePlus is trained and evaluated on the Aalto University Keystroke Dataset, which consists of keystroke data collected from 168,000 participants using physical keyboards (desktop or laptop). The keystroke events were recorded with millisecond precision using JavaScript functions, capturing both key press (keydown) and release (keyup) events in Coordinated Universal Time (UTC) format.

2. Data Preprocessing

Each keystroke sequence was processed to extract four temporal features, capturing the time relationships between consecutive keystrokes:

  1. Hold Latency (HL): Time elapsed between pressing and releasing the same key.
  2. Inter-key Latency (IL): Time between the release of one key and the press of the next key.
  3. Press Latency (PL): Time between two consecutive key press events.
  4. Release Latency (RL): Time between two consecutive key release events.

These features were computed for each keystroke pair and converted from milliseconds to seconds to ensure consistency in time representation. Unlike previous studies that treated keycodes as numerical values or categorical features, we treated keycodes as embeddings. Each unique key was mapped to a learnable dense representation, allowing the model to capture meaningful relationships between frequently used key pairs.

3. TypePlus Architecture

The proposed TypePlus architecture is designed as an improvement for TypeNet architecture. The model consists of multiple layers, including Long ShortTerm Memory (LSTM) networks, attention pooling, and a projection layer that generates meaningful representations for each user’s keystroke dynamics. The figure below illustrates the overall architecture of TypePlus. TypePlus Architecture: Verification scheme based from two users i and j. The input x is a time series with shape M × 5 (keystrokes × keystroke features) and the output f(x) is an embedding vector with shape 1 × 128. t is a decision threshold.

After preprocessing, we obtain five features: keycodes and four temporal feature series. The transformed input is then fed into a TypePlus embedding model comprising several key components. First, an Embedding Layer maps input keycodes into dense vector representations, allowing the model to more effectively capture spatial relationships between different keycodes. For sequences with lengths N > M, we truncate the input to fit within M. Conversely, when N < M, we apply zero-padding at the end of the sequence until it reaches M. The Masking mechanism ensures that these padded values are ignored during training. The model employs a hierarchical LSTM architecture, consisting of two sequential bidirectional LSTM layers, each with 128 hidden units per direction (256 total per layer due to bidirectionality).

A critical component of the architecture is the Attention Pooling mechanism, which enables the model to dynamically focus on the most relevant keystroke features. This mechanism is implemented using a fully connected attention layer that assigns importance scores to each time step in the sequence. Given an input sequence x, the attention scores αt are computed as follows:

α_t = softmax(W_a * x_t)

These scores are then applied to the LSTM outputs to compute a weighted sum:

z = ∑(α_t * x_t)

Verification Scheme

The Euclidean distance between keystroke embeddings is calculated as:

d = || f(x_i) - f(x_j) ||_2

where a lower distance indicates a higher similarity. A threshold t is applied for classification:

  • If ( d < t )Genuine User
  • If ( d ≥ t )Impostor

Results

In this section, we present the empirical findings of our keystroke authentication experiments under four primary loss functions: Contrastive Loss, Euclidean Triplet Loss, Manhattan Triplet Loss, and Filtered Triplet Loss. We evaluate two architectures: – TypeNet: The baseline recurrent model for keystroke dynamics. – TypePlus: The enhanced variant that introduces a learnable embedding layer for keycodes, bidirectionality, and attention-based pooling.

Model Loss Function EER (%)
TypeNet Contrastive Loss 10.56%
TypeNet Manhattan Triplet Loss 6.49%
TypeNet Euclidean Triplet Loss 6.28%
TypeNet Filtered Triplet Loss 6.52%
TypePlus Contrastive Loss 5.50%
TypePlus Manhattan Triplet Loss 2.93%
TypePlus Euclidean Triplet Loss 2.90%
TypePlus Filtered Triplet Loss 2.86%

TypePlus significantly outperforms previous architectures, demonstrating the effectiveness of attention pooling and loss function optimization.

About

TypePlus is a lightweight deep learning model for free-text keystroke authentication, evaluated across multiple loss functions, including a Filtered Mean Triplet Loss. It achieves state-of-the-art performance with a 2.86% Equal Error Rate (EER) on the Aalto University Keystroke Dataset, demonstrating its effectiveness in keystroke biometrics.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published