It now supports fine-tuning, and the inference cost is the same as the base model! <coughs LORA adopters> π€π€
So the base model must be expensive? πΈ For the base model, the input price is reduced by 78% to $0.075/1 million tokens and the output price by 71% to $0.3/1 million tokens. ππ΅
But is it any good? π€·ββοΈ On the LLM Hallucination Index, Gemini 1.5 Flash achieved great context adherence scores of 0.94, 1, and 0.92 across short, medium, and long contexts. ππ―
Google has finally given a model that is free to tune and offers an excellent balance between performance and cost. βοΈπ