--- license: apache-2.0 base_model: Qwen/Qwen2.5-Coder-3B-Instruct tags: - n8n - workflow - json - code - fine-tuned - merged language: - fr - en pipeline_tag: text-generation --- # Qwen Coder 3B - N8N Workflow Generator (Merged) This is the **merged** version of the LoRA fine-tuned model for generating N8N workflow JSON. ## Model Details - **Base Model**: Qwen/Qwen2.5-Coder-3B-Instruct - **Fine-tuned adapter**: eclaude/qwen-coder-3b-n8n-sft - **Training**: SFT on 8,782 N8N workflow examples - **Task**: Generate valid N8N workflow JSON from natural language prompts (French) ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("eclaude/qwen-coder-3b-n8n-merged") tokenizer = AutoTokenizer.from_pretrained("eclaude/qwen-coder-3b-n8n-merged") prompt = "Crée un workflow qui récupère des données d'une API et les envoie sur Slack" inputs = tokenizer(prompt, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=2048) print(tokenizer.decode(outputs[0])) ``` ## Training Data - Dataset: [eclaude/n8n-workflows-sft](https://huggingface.co/datasets/eclaude/n8n-workflows-sft) - 8,782 training samples - 1,197 evaluation samples ## Metrics - Training loss: 1.04 - Eval loss: 1.02 - Token accuracy: 73%