Any performance benchmarks compared to other LLMs? Also, any performance increases on the orig Falcon model in inference speed?<p>We ditched most of our focus on Falcon 40B after Llama 2 70B came out, both the tokens per sec and quality of results are not even close.
so its open source? Where is the link to the weights?<p>The article was mildly annoying because of many underlines sentences and words that looked like links.