“Unusual Perspectives on AI” – Computer and System Architecture Unraveled Event Five

The fifth CaSA, Computer and System Architecture Unraveled, meetup took place on January 30. We finally gave in and joined the AI hype train, resulting in an event with a somewhat different audience and different discussions. More society and applications, less computer architecture. Our two presenters were Håkan Zeffer from SambaNova Systems and Björn Forsberg from RI.SE (doing his second CaSA presentation!). Håkan talked about the architecture of the Sambanova AI processors, and Björn about AI compilers.

Continue reading ““Unusual Perspectives on AI” – Computer and System Architecture Unraveled Event Five”

(Local) AI, Please Reason about Code

“Reasoning” models have become popular as a way to expand the capabilities of large language models (LLMs). Such models take more time “considering” a prompt and iterating it through the model several times, with the goal of mimicking how a human might go about solving a problem by breaking it down into steps. I tried the reasoning QwQ model on the coding problems from my previous blog posts (1,2). Quite funny and elucidating; I will quote the replies in full as they are worth reading.

Continue reading “(Local) AI, Please Reason about Code”

(Local) AI, Please Write some Code

My previous blog post in this series tested the ability of a range of large language models to analyze a piece of C code and determine what a mystery function did. That was interesting and entertaining, but possibly not a particularly “fair” test of the models’ capabilities. Most of time, I think people use “AI” to help write code, not to understand some tricky piece of algorithmic code. Thus, I turn the problem around and ask the models to write code for the algorithm I previously asked them to analyze.

Continue reading “(Local) AI, Please Write some Code”

(Local) AI, Please Explain This Code

Continuing my exploration of what a local AI model can do, I decided to test them on the task of code analysis. It would be so nice to have an AI model that is tuned and trained on a particular tool or programming system, and that can be distributed for users to run on their own on their local machine, server, or cloud VM. Avoiding the need to run and charge for a custom cloud service and ensuring confidentiality and availability.

Updated 2024-12-12 with Llama-3.3-70B

Continue reading “(Local) AI, Please Explain This Code”

More Exploration of (Local) AI Models

In my previous blog post about the Intel AI Playground, I tested it by asking it to draw cars. In this post, I share some more exploration of these local AI models and their limitations. Turns out that cars are easy, other things not so much…

Continue reading “More Exploration of (Local) AI Models”

Delivering AI-Based Solutions is not Always Easy

One of the nice properties of delivering software that users install on their own machines is that once the software has been built and shipped, the cost of running it is handed over to the user. The cost per installation and per user is minimal in terms of compute load on the developing company. Of course there are costs for things like support, but that is different. However, having the customer provide the compute resources is not necessarily that easy when it comes to AI-based setups.

Continue reading “Delivering AI-Based Solutions is not Always Easy”

IBM WatsonX-AI, Kista Tech Tuesday

I attended a short Tech Tuesday morning session at IBM here in Kista, Sweden, where IBM presented their WatsonX-AI and related technologies. Tech Tuesdays is a monthly technical event organized by Kista Science City, where companies in Kista present some aspect of their technology in a 30-minute session. IBM managed to get an impressive amount of content into that time!

Continue reading “IBM WatsonX-AI, Kista Tech Tuesday”

RI.SE AI Day – More on LLMs (and some)

The Swedish research institute RI.SE hosted an “Artificial Intelligence and Computer Science day” (AI and CS day) last week. RI.SE has a long tradition of hosting interesting open houses, both as RI.SE and in their previous guide as SiCS. The day was a mix of organized talks in the morning, and an open house where RI.SE researchers showed off their work in the afternoon. Most of the AI discussions were related to large language models (LLMs), but not all. I got some new insights about LLMs in general and using LLMs for coding in particular.

Continue reading “RI.SE AI Day – More on LLMs (and some)”

ChatGPT and Legal

In previous three blog posts (1,2,3) about ChatGPT in particular and large language models in general, I touched on what they can do, what they cannot do, what they seem not to do, how they fall down in funny ways, and why I think they are fundamentally flawed for many applications. There is one more aspect left to consider – the legal and licensing side. I am not a lawyer, I am not an expert, but it seems obvious that there is a huge problem. There are also clear questions about business morals and what the right thing to do would be. I also doubt the business viability of LLMs in the way they are currently trained.

Continue reading “ChatGPT and Legal”

ChatGPT and Critique

ChatGPT and other transformer-based models like Dall-E are technologically very impressive. They do things that seemed totally impossible just a few years ago. However, they are not really generally intelligent, and there are innumerable problems with how they work, what they do, what people think they do, ethics, and legal and licensing issues. This is my third post about ChatGPT, where I present my critique of and reflections on the technology. The previous posts were about ChatGPT and Simics and Coding using ChatGPT.

Continue reading “ChatGPT and Critique”

ChatGPT and Code

In my previous blog post about ChatGPT and Simics, I tested it on its knowledge and abilities with a fairly niche subject. Not unsurprisingly it did not do all that well. However, one area where ChatGPT appears to really work well is when dealing with program code. This seems more practically useful as well, especially as a generator of starting points and boiler-plate code. It can also sometimes do a decent job explaining code, subject to quite common bizarre mistakes and errors. Update: Part 3, a critique of ChatGPT has been published.

Continue reading “ChatGPT and Code”

ChatGPT and Simics

It is an understatement to say that ChatGPT has been a hot topic since it was launched a few months back. Everyone seems to be seeing what it can do in their favorite domain, so I had to try it on what I work with, Simics and virtual platforms. The results did not live up to the hype some people think the technology deserves, but it was very impressive and a little scary nevertheless. This is the first post in what looks like it will be a series about ChatGPT. Update: Part 2, ChatGPT and Code, is now out. Update to the update: Part 3, a critique of ChatGPT has been published.

Continue reading “ChatGPT and Simics”