--- library_name: transformers license: other base_model: meta-llama/Meta-Llama-3.1-8B tags: - llama-factory - full - generated_from_trainer model-index: - name: stabletoolbench_train_160k results: [] --- # *From Word to World*: Can Large Language Models be Implicit Text-based World Models? [![arXiv](https://img.shields.io/badge/arXiv-2512.18832-b31b1b?logo=arXiv)](https://arxiv.org/abs/2512.18832) [![Blog](https://img.shields.io/badge/Blog-Post-blue?logo=rss&logoColor=white)](https://macaron.im/mindlab/research/how-world-models-unlock-scalable-agentic-rl) [![HF Paper](https://img.shields.io/badge/Paper-HuggingFace-yellow?logo=huggingface&logoColor=white)](https://huggingface.co/papers/2512.18832) [![Models](https://img.shields.io/badge/Models-HuggingFace-yellow?logo=huggingface&logoColor=white)](https://huggingface.co/collections/X1AOX1A/llm-as-world-models) [![Dataset](https://img.shields.io/badge/Dataset-HuggingFace-yellow?logo=huggingface&logoColor=white)](https://huggingface.co/datasets/X1AOX1A/LLMasWorldModels) # stabletoolbench_train_160k This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) on the stabletoolbench_train_175183 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - gradient_accumulation_steps: 8 - total_train_batch_size: 128 - total_eval_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: constant_with_warmup - lr_scheduler_warmup_steps: 10 - num_epochs: 5.0 ### Training results ### Framework versions - Transformers 4.52.4 - Pytorch 2.9.0+cu128 - Datasets 3.6.0 - Tokenizers 0.21.1