I hate seeing these guys succeed because everyone of their successes is a new day that AI becomes less accessible to the average person and more locked behind their APIs.
I see this statement a lot and have no idea how people come to this conclusion. I have a beefy 16k$ workstation with 2 4090s and I could barely run the LLAMA 65B model at a very slow pace. Let us say we do have the model weights to GPT-4 and GPT3.5, me as the average consumer I don't know how this helps me in any way. I need to shell at least 25k (possibly much more for GPT-4) before I can run these models for even inference, and even then it will be a slow, unpolished experience.
On the other hand OpenAI’s API makes things blazingly fast and dirt cheap to the average consumer. It honestly does feel like they have enabled the power of AI to be accessible to anyone with a laptop. If that requires fending off competition from Behemoths like Google, Meta by not releasing model weights then so be it. This critique would be more apt to Nvidia who are artificially increasing datacenter GPU prices thus pricing out the average consumer. OpenAI is doing the opposite.
I have been thinking about trying to do LoRA style fine tuning of Flacon-40b or Falcon-7b on RunPod. The new OpenAI 16k context and functions thinking made me lose the urge to get into that. Was questionable whether it could really write code consistently anyway even if very well fine tuned.
But at least that is something that can be attempted without $25k.
How does it compare to GPT 3.5, or 4? I mean if you ask the same questions. Is it usable at all?
I tried the models that work with 4090 and they were completely useless for anything practical (code questions, etc.). Curiosities sure, but on Eliza level.
the one that I used for GPT-4 and the local ones was a bit obscure:
"how to configure internal pull-up resistor on PCA9557 from NXP in firmware"
the GPT4 would give a paragraph of
> The PCA9557 from NXP is an 8-bit I/O expander with an I2C interface. This device does not have an internal pull-up resistor on its I/O pins, but it does have software programmable input and output states.
and then write a somewhat meaningful code. the local LLMs failed even at the paragraph stage
This is exactly my problem. They are doing quite well, and closing the door behind them. Open AI isn't your friend and reserves the right to screw you down the line.
Thing is, as long as the field is growing in capabilities as fast as it is, there isn't going to be any kind of "democratizing" for an average person, or even average developer. Anything you or me can come up with to do with LLM, some company or startup will do better, and they'll have people working full-time to productize it.
Maybe it's FOMO and depression, but with $dayjob and family, I don't feel like there's any chance of doing anything useful with LLMs. Not when MS is about to integrate GPT-4 with Windows itself. Not when AI models scale superlinearly with amount of money you can throw at them. I mean, it's cool that some LLAMA model can run on a PC, provided it's beefy enough. I can afford one. Joe Random Startup Developing Shitty Fully Integrated SaaS Experience can afford 100 of them, plus an equivalent of 1000 of them in the cloud. Etc.