|
|
--- |
|
|
license: cc-by-4.0 |
|
|
tags: |
|
|
- malware |
|
|
- cybersecurity |
|
|
- ATT&CK |
|
|
- MBC |
|
|
- pe-files |
|
|
- elf-files |
|
|
- binary-classification |
|
|
- tabular-data |
|
|
- threat-intelligence |
|
|
- digital-forensics |
|
|
- reverse-engineering |
|
|
- incident-response |
|
|
- security-telemetry |
|
|
- ai-security |
|
|
- security-ml |
|
|
- mitre-attack |
|
|
- mitre-mbc |
|
|
- windows |
|
|
- linux |
|
|
- executable-files |
|
|
- static-analysis |
|
|
- behavioral-analysis |
|
|
- classification |
|
|
- anomaly-detection |
|
|
- intrusion-detection |
|
|
- explainable-ai |
|
|
- model-evaluation |
|
|
- benchmarking |
|
|
- training |
|
|
- evaluation |
|
|
- research |
|
|
- education |
|
|
- teaching |
|
|
pretty_name: traceix-ai-telemetry |
|
|
--- |
|
|
|
|
|
# Traceix AI Security Telemetry |
|
|
|
|
|
Each dataset is a JSONL file where **each line describes a single file analyzed by [Traceix](https://traceix.com)**. For every file you get: |
|
|
|
|
|
* **`file_capabilities`** – high-level behaviors and capabilities (CAPA-style + mapped to ATT&CK and MBC tags like `Execution/T1129`, `Discovery/T1083`, etc.). |
|
|
* **`file_exif_data`** – parsed EXIF metadata (file size, type, timestamps, company/product info, subsystem, linker/OS versions, etc.). |
|
|
* **`model_classification_info`** – [Traceix](https://traceix.com) model verdict (`safe` / `malicious`), classification timestamp, and inference latency in seconds. |
|
|
* **`decrypted_training_data`** – numeric feature vector actually used for training/inference (PE header fields, section statistics, imports/resources counts, entropy stats, etc.). |
|
|
* **`metadata`** – model version and accuracy, upload metadata (timestamp, SHA-256, license), and payment information (THRT amount, Solana transaction hash + explorer URL, price at time of payment). |
|
|
|
|
|
All records are focused on **malware analysis** and are stored in **JSONL format**. |
|
|
Datasets are **automatically exported by [Traceix](https://traceix.com) on a monthly schedule** and published **as-is** under the **CC BY 4.0** license. |
|
|
|
|
|
You can quickly load and sanity-check any monthly corpus using: |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
import pandas as pd |
|
|
from sklearn.linear_model import LogisticRegression |
|
|
from sklearn.model_selection import train_test_split |
|
|
|
|
|
|
|
|
# Load the Traceix telemetry dataset |
|
|
ds = load_dataset( |
|
|
"PerkinsFund/traceix-ai-security-telemetry", |
|
|
data_files="traceix-telemetry-corpus-2025-12.jsonl", # Or whatever month you want |
|
|
split="train", |
|
|
) |
|
|
|
|
|
# We will need to flatten nested JSON into columns |
|
|
df = ds.to_pandas() |
|
|
df_flat = pd.json_normalize(df.to_dict(orient="records")) |
|
|
|
|
|
# Define the features and label based on schema |
|
|
feature_cols = [ |
|
|
"decrypted_training_data.SizeOfCode", |
|
|
"decrypted_training_data.SectionsMeanEntropy", |
|
|
"decrypted_training_data.ImportsNb", |
|
|
] |
|
|
|
|
|
label_col = "model_classification_info.identified_class" |
|
|
|
|
|
# We don't have to but we will drop the rows with missing data |
|
|
df_flat = df_flat.dropna(subset=feature_cols + [label_col]) |
|
|
|
|
|
X = df_flat[feature_cols].values |
|
|
y = (df_flat[label_col] == "malicious").astype(int) |
|
|
|
|
|
# Start the training and test split |
|
|
X_train, X_test, y_train, y_test = train_test_split( |
|
|
X, y, test_size=0.2, random_state=42 |
|
|
) |
|
|
|
|
|
# Test the basic file classifier |
|
|
clf = LogisticRegression(max_iter=1000) |
|
|
clf.fit(X_train, y_train) |
|
|
|
|
|
print("Test accuracy:", clf.score(X_test, y_test)) |
|
|
``` |