Experiments to evaluate multi-lingual capability of Gemini
Pranav Mani
Many parts of the world are multi-lingual
Humans are incredibly efficient at processing “long contexts” of multilingual text. They can detect languages, effectively reason and recall information with extreme accuracy.
Question
Is the language detection capability of Gemini 1.5 robust to handle large amount of multilingual text?
Collecting dataset
I scraped random articles per language from Wikipedia upto 100k words. The script used for scraping articles is also shared.
Why random order? Switching topics randomly from one article to another effectively “stresses” the long context capabilities of Gemini.
The datasets are classified per language. Multi-lingual datasets�are generated using single language datasets by concatenating�them.
Experiments and results
Without Context Caching:
Add on experiment: Simulating a bilingual conversation with 2 Gemini APIs.
�
Takeaways