I tested this out on my workload ( SRE/Devops/C#/Golang/C++ ). it started responding about non-sense on a simple write me boto python script that changes x ,y,z value.<p>Then I tried other questions in my past to compare... However, I believe the engineer who did the LLM, just used the questions in benchmarks.<p>One instance after a hour of use ( I stopped then ) it answered one question with 4 different programming languages, and answers that was no way related to the question.