A Small Journey in AI
2024-05-28
“I love you very much”
V1 | V2 | V3 | V4 | … | V512 | |
---|---|---|---|---|---|---|
I | 0.60 | 0.00 | 1.39 | -0.60 | … | -0.64 |
love | 0.00 | -2.12 | -1.37 | -1.10 | … | -1.20 |
you | -1.55 | -0.24 | 1.90 | 0.48 | … | -0.76 |
very | 0.27 | -0.81 | -2.46 | 1.36 | … | 0.62 |
much | 2.33 | 0.59 | -1.26 | -0.76 | … | 0.28 |
“If I should think of love I’d think of you, your arms uplifted, Tying your hair in plaits above, The lyre shape of your arms and shoulders, The soft curve of your winding head. No melody is sweeter, nor could Orpheus So have bewitched. I think of this, And all my universe becomes perfection. But were you in my arms, dear love, The happiness would take my breath away, No thought could match that ecstasy, No song encompass it, no other worlds. If I should think of love, I’d think of you.”
– Shakespeare
“If I should think of love I’d think of you, your arms uplifted, Tying your hair in plaits above, The lyre shape of your arms and shoulders, The soft curve of your winding head. No melody is sweeter, nor could Orpheus So have bewitched. I think of this, And all my universe becomes perfection. But were you in my arms, dear love, The happiness would take my breath away, No thought could match that ecstasy, No song encompass it, no other worlds. If I should think of love, I’d think of you.”
– Shakespeare
Content generation
In March 2021, GPT-3 was typing 3.1 million words per minute, non-stop, 24×7. With the general availability of the model, I expect that number is a lot higher now… (Nov/2021).
Hardware
The supercomputer developed for OpenAI (May 2020) is a single system with more than 285,000 CPU cores, 10,000 GPUs 1 and 400 gigabits per second of network connectivity for each GPU server.
– https://blogs.microsoft.com/ai/openai-azure-supercomputer/
Training time
Training GPT-3 with 175 billion parameters would require approximately 288 years with a single V100 NVIDIA GPU.
– https://arxiv.org/pdf/2104.04473.pdf
Understanding
Despite the impending widespread deployment of foundation models, we currently lack a clear understanding of how they work, when they fail, and what they are even capable of due to their emergent properties.
– https://arxiv.org/pdf/2108.07258
We’re going to deploy a quantized model of Phi-3 which can run on consumer hardware
Because we can
Running your own language model enables you to be productive, safe, compliant, anonymous and autonomous. Also without an internet connection..
What it is
Michael Green, CEO Desupervised mike@desupervised.io +4531766142