Keep the queries short, no more than 3-5 words long. Focus on a single task and a single species, and include the species name in the query. This allows a more focused context to be presented to the LLM. Here are a few examples to get you started:
- What is Saigona baiseensis?
- Describe Saigona sinicola
- Describe Carvalhoma malcolmae
- Does Rhinolophus sinicus live in caves?
- Where does Laephotis botswanae roost?
- What distinguishes Choanolaimus sparsiporus?
- What is the etymology of Gammanema lunatum?
- Describe Amnestus sinuosus
- What distinguishes Cynodon gibbus from other cofamilial genera?
About
Zai is the Zenodeo ai providing a Q&A access to the ~1M treatments extracted by Plazi and made available on TreatmentBank and BLR on Zenodo.
Zai responds to simple and focused questions with easy-to-understand summaries or pointed answers. It cannot (yet) respond to questions that span many documents, and is best at extracting answers from a single treatment that can be pin-pointed with a full-text search.
Zai is also the Chinese verb ε¨ (zΓ i) for describing existence in a location, similar to how we say in English, "to be at" or "to be in."
Semantic Cache
We use the Supabase/gte-small model for generating vector embeddings of queries.
βββββββββββββ
βwhat is theβ
ββββββββββββββββββββββββββββββββΆβmeaning of ββββ
β βlife? β β
asks βββββββββββββ β
β β β
β β β β β β β β β β β β β β β βββ β β β βΌ
β semantic cache β ββ
ββββββ β βββββββββββ ββββββββΌβββββ β
β\O/ β βsearch β βconvert β ββ
β | β 42 β βfor exactβ βquery to β β
β/ \ βββββββββfoundββ€key in βββββunique β ββ
βuserβ β βcache β βcache key β β
ββββββ ββββββ¬βββββ βββββββββββββ ββ
β² β β β
β not found ββ
β β β β
β ββββββΌβββββ ββ
β β βconvert β β
β βto β ββ
β β βembeddingβββββββββββββββββββββ
β βvector β β
β β βββββββββββ
β β
β β ββββββββββββββββββββββ
β βsearch for query β β
β β βwith highest cosine β
βββββββββββfoundββsimilarity above a β β
β βpreset threshold β
ββββββββββββββββββββββ β
β β
β β β β β β β β β βΌ β β β β β β β β β β
no
β
βΌ
How this works
βββββββββββββ
βwhat is theβ
ββββββββββββββΆβmeaning of β
β βlife? β
asks βββββββββββββ
β β
β β β β β β βββ β β β ββ β β β β β β β β β
ββββββ β semantic cache β
β\O/ β β βββββββΌββββββ
β | β yes βis the β ββββββββ β
β/ \ ββββ42βΌβββββanswer in βββββββββββΆβcache β
βuserβ βcache? β βββββ²βββ β
ββββββ β βββββββββββββ β
no β β
β β β β β βββ β β β ββ β β β β βββ β β β
β β
ββββββββββΌββββββββ β
βremove stopwordsβ βββββββ΄βββββ
β"what", "is", β β response β
β"the", and "of" β βββββββ²βββββ
ββββββββββ¬ββββββββ β
β βββββββ΄βββββ
ββββββββββΌββββββββ β LLM β
βfull text searchβ βββββββ²βββββ
βof "meaning" andβ β
β"life" β β
ββββββββββ¬ββββββββ ββββββββββ΄ββββββββ
β βquestion: "what β
β βis the meaning β
β βof life?" β
β top ββΆβ β
β ranked βcontext: full β
β paper βtext of the β
ββββββββΌββββββ β βpaper β
βpapers β β ββββββββββββββββββ
βranked by ββββ
βrelevance β
ββββββββββββββ
We are testing various LLMs including Alibaba's Qwen 3:0.6b, Meta's Llama 3.2:1b and 3b, and Google's Gemma 3 to generate answers to questions based on the context provided. The context is chosen from the full text search result of the relevant tokens in your query ranked by their relevance (BM25 score). The generated answers are stored in a semantic cache for fast retrieval for subsequent queries.