I run AI applications all the time in exactly those situations. The models range from 2GB (vector models) 30GB (small LLMs) to 100GB (medium LLMs).
None of those fit in 4MB of cache (the per-core on this part), or 1GB (the aggregate cache).
What AI models are you actually talking about? Do you mean old-school ML stuff, like decision trees or high dimensional indexes? No one I know calls those "AI", which is generally reserved for big-ish neural networks.