`;return o},i=(function(){let g={"809":{"id":809,"name":"Claude 3.5 Haiku","model_name":"Claude 3.5 Haiku","desc":"This is the next generation of Anthropic's fastest model. For a similar speed to Claude 3 Haiku, Claude 3.5 Haiku improves across every skill set and surpasses even Claude 3 Opus, the largest model in their previous generation, on many intelligence benchmarks. Claude 3.5 Haiku is particularly strong on coding tasks. It also features low latency, improved instruction following, and more accurate tool use.","desc_short":"
This is the next generation of Anthropic's fastest model. For a similar speed to Claude 3 Haiku, Claude 3.5 Haiku improves across every skill set and surpasses even Claude 3 Opus, the largest model in their previous generation, on many intelligence benchmarks. Claude 3.5 Haiku is particularly strong on coding tasks. It also features low latency, improved instruction following, and more accurate tool use.
","desc_more":"","link":"claude-3.5-haiku","provider":"Official","developer":"Anthropic","tpm":60,"image":"https://hbcdn01.hotbot.com/avatar/809.jpg","model":189,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":2,"ph":"","labels_full":["Official","Premium","Large Context","Claude","Unmodified","Coding","Creative Writing"],"labels":["Official","Premium"],"wel":"","asktext":"","short_name":"Claude 3.5 Haiku","set":{"temperature":1}},"849":{"id":849,"name":"DeepSeek R1","model_name":"DeepSeek R1","desc":"First-generation reasoning model from DeepSeek. Open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass.","desc_short":"
First-generation reasoning model from DeepSeek. Open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass.
","desc_more":"","link":"deepseek-r1","provider":"Official","developer":"DeepSeek","tpm":28,"image":"https://hbcdn01.hotbot.com/avatar/849.jpg","model":209,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":1,"ph":"","labels_full":["Official","Premium","Uncensored","Large Context","DeepSeek","Unmodified","Coding","Advanced Reasoning","671B Parameters"],"labels":["Official","Premium"],"wel":"","asktext":"","short_name":"DeepSeek R1","set":{"temperature":1}},"289":{"id":289,"name":"MythoMax 13B","model_name":"MythoMax 13B","desc":"One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay.","desc_short":"
One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay.
","desc_more":"","link":"mythomax-13b","provider":"Official","developer":"gryphe","tpm":2,"image":"hotbot.png","model":117,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":0,"ph":"","labels_full":["Official","Unlimited","Uncensored","Unmodified","13B Parameters"],"labels":["Official"],"wel":"","asktext":"","short_name":"MythoMax 13B","set":{"temperature":1}},"321":{"id":321,"name":"OpenChat 3.5 7B","model_name":"OpenChat 3.5 7B","desc":"OpenChat 7B is a library of open-source language models, fine-tuned with C-RLFT (Conditioned Reinforcement Learning Fine-Tuning) - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels.","desc_short":"
OpenChat 7B is a library of open-source language models, fine-tuned with C-RLFT (Conditioned Reinforcement Learning Fine-Tuning) - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels.
","desc_more":"","link":"openchat-3.5-7b","provider":"Official","developer":"OpenChat","tpm":2,"image":"https://hbcdn01.hotbot.com/avatar/321.jpg","model":149,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":0,"ph":"","labels_full":["Official","Unlimited","Uncensored","OpenChat","Unmodified","7B Parameters"],"labels":["Official"],"wel":"","asktext":"","short_name":"OpenChat 3.5 7B","set":{"temperature":1}},"137":{"id":137,"name":"OpenAI GPT-3.5 Turbo","model_name":"OpenAI GPT-3.5 Turbo","desc":"OpenAI GPT-3.5 Turbo is a fast, inexpensive model for simple tasks developed by OpenAI. Capable of understanding and generating natural language or code, it has been optimized for chat.","desc_short":"
OpenAI GPT-3.5 Turbo is a fast, inexpensive model for simple tasks developed by OpenAI. Capable of understanding and generating natural language or code, it has been optimized for chat.
","desc_more":"","link":"openai-gpt-3.5-turbo","provider":"Official","developer":"OpenAI","tpm":20,"image":"https://hbcdn01.hotbot.com/avatar/137.jpg","model":5,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":1,"ph":"","labels_full":["Official","Premium","ChatGPT","OpenAI","Unmodified","Coding"],"labels":["Official","Premium"],"wel":"","asktext":"","short_name":"OpenAI GPT-3.5 Turbo","set":{"temperature":1}},"201":{"id":201,"name":"OpenAI GPT-4o Mini","model_name":"OpenAI GPT-4o Mini","desc":"GPT-4o Mini is OpenAI's most advanced model in the small models category. It is multimodal (accepting text or image inputs and outputting text), and has higher intelligence than gpt-3.5-turbo but is just as fast. It is meant to be used for smaller tasks, including vision tasks.","desc_short":"
GPT-4o Mini is OpenAI's most advanced model in the small models category. It is multimodal (accepting text or image inputs and outputting text), and has higher intelligence than gpt-3.5-turbo but is just as fast. It is meant to be used for smaller tasks, including vision tasks.
","desc_more":"","link":"openai-gpt-4o-mini","provider":"Official","developer":"OpenAI","tpm":8,"image":"https://hbcdn01.hotbot.com/avatar/201.jpg","model":89,"sys":"","read_image":1,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":0,"ph":"","labels_full":["Official","Unlimited","Large Context","OpenAI","ChatGPT","Unmodified","Image Input","Coding","Creative Writing"],"labels":["Official"],"wel":"","asktext":"","short_name":"Free ChatGPT","set":{"temperature":1}},"205":{"id":205,"name":"Meta Llama3.1 8b","model_name":"Meta Llama3.1 8b","desc":"Llama 3.1 is a group of open-source instruction-tuned models from Meta. These multilingual models have a context length of 128K, state-of-the-art tool use, and strong reasoning capabilities. The 8B variant is a light-weight, ultra-fast model that can run anywhere.","desc_short":"
Llama 3.1 is a group of open-source instruction-tuned models from Meta. These multilingual models have a context length of 128K, state-of-the-art tool use, and strong reasoning capabilities. The 8B variant is a light-weight, ultra-fast model that can run anywhere.
","desc_more":"","link":"meta-llama3.1-8b","provider":"Official","developer":"Meta","tpm":2,"image":"https://hbcdn01.hotbot.com/avatar/205.jpg","model":93,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":0,"ph":"","labels_full":["Official","Unlimited","Meta","Llama","Unmodified","8B Parameters","Coding"],"labels":["Official"],"wel":"","asktext":"","short_name":"Meta Llama3.1 8b","set":{"temperature":1}},"145":{"id":145,"name":"OpenAI GPT-4 Turbo","model_name":"OpenAI GPT-4 Turbo","desc":"GPT-4 is a large multimodal model (accepting text or image inputs and outputting text) that can solve difficult problems with greater accuracy than any of OpenAI's previous models, and the turbo variant is optimized for quick responses and natural language chat.","desc_short":"
GPT-4 is a large multimodal model (accepting text or image inputs and outputting text) that can solve difficult problems with greater accuracy than any of OpenAI's previous models, and the turbo variant is optimized for quick responses and natural language chat.
","desc_more":"","link":"openai-gpt-4-turbo","provider":"Official","developer":"OpenAI","tpm":20,"image":"https://hbcdn01.hotbot.com/avatar/145.jpg","model":13,"sys":"","read_image":0,"prem":false,"can_at":1,"no_comp":0,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":1,"ph":"","labels_full":["Official","Premium","Large Context","OpenAI","ChatGPT","Unmodified","Coding","Creative Writing"],"labels":["Official","Premium"],"wel":"","asktext":"","short_name":"OpenAI GPT-4 Turbo","set":{"temperature":1}},"1":{"id":1,"name":"HotBot Assistant","model_name":"HotBot Assistant","desc":"HotBot Assistant is a comprehensive chat companion that can help you with a number of tasks based on how you talk to it.","desc_short":"
HotBot Assistant is a comprehensive chat companion that can help you with a number of tasks based on how you talk to it.
","desc_more":"","link":"hotbot-assistant","provider":"Official","developer":"HotBot","tpm":14,"image":"https://hbcdn01.hotbot.com/avatar/1.jpg","model":201,"sys":"","read_image":1,"prem":false,"can_at":0,"no_comp":1,"search":0,"no_search":0,"no_audio":0,"no_attach":0,"wizard":[],"repeat_wizard":0,"per_message":0,"level":0,"ph":"","labels_full":["Official","Unlimited","Image Generation","Web Search","Coding","70B Parameters","Creative Writing"],"labels":["Official"],"wel":"","asktext":"","short_name":"Assistant","set":{"temperature":1}}},b={},d=localStorage,n="hba-botcache",l=()=>{let g={},b=d[n];try{let l=Date.now()-6e5;b&&$.each(JSON.parse(b),(b,d)=>{if(d.cacheTS>l)g[b]=d})}catch(a){}return g},a=g=>{let b=l();for(let a of g){b[a.id]=a;if(!a.cacheTS)a.cacheTS=Date.now()}d[n]=JSON.stringify(b)};$.each(l(),(b,d)=>{if(!g[b])g[b]=d});async function e(d){let n=[],l={};for(let h of(d instanceof Array)?d:[d])if(h=parseInt(h)){if(b[h]&&!g[h])await b[h];if(g[h])l[h]=g[h];else if(n.indexOf(h)<0)n.push(h)}if(n.length){let i=$.post("/lookup-bots.json",{bot_id:n.join(",")});for(let h of n)b[h]=i;let o=await i;if(o.code==0){a(o.data);for(let h of o.data)g[h.id]=h,l[h.id]=h}for(let h of n)delete b[h]}return l}return{cache:g,getBots:e,getBot:async g=>g?(await e(g))[g]:null}})();$('