ai

Resources

Large Language Model (LLM)

https://www.youtube.com/watch?v=zjkBMFhNj_g

Two files - llama-2-70b model

Screenshot 2024-02-28 at 6.37.09 PM.png

Model training is a lot more intensive than model inference

Model training is like the compression of the internet.

Screenshot 2024-02-28 at 6.40.38 PM.png

Cost about 2 million dollars, and about 12 days to train the 70b model

This is the lossy compression of the internet. We don't have an identical copy of it.

Neural network - Predict the next word in the sequence

Screenshot 2024-02-28 at 6.42.42 PM.png

next word prediction

the network dreams internet documents

Screenshot 2024-03-05 at 5.35.04 PM.png

neural network built by multiple steps of optimization
can measure the output in different situations

Screenshot 2024-03-05 at 5.37.17 PM.png

The assistant model

Screenshot 2024-03-05 at 5.40.51 PM.png

Fine tuning is a lot cheaper

Screenshot 2024-03-05 at 5.44.33 PM.png

Screenshot 2024-03-05 at 5.48.43 PM.png

What accuracy are you going to achieve?

Multimodalities

Thinking fast vs thinking slow: instinctive vs slower
GenAI is a system 1 -- words enter in a sequence

Screenshot 2024-03-05 at 6.00.03 PM.png

AlphaGo

Custom LLMs

Screenshot 2024-03-05 at 6.05.07 PM.png

Augmented generation

Thinking for a long time using system 2

Kernel - an operating system

Screenshot 2024-03-05 at 6.08.34 PM.png

Jailbreak attacks

Screenshot 2024-03-05 at 6.11.00 PM.png

Base64 jailbreak
The LLM is fluent in Base64

Screenshot 2024-03-05 at 6.14.58 PM.png
Screenshot 2024-03-05 at 6.19.42 PM.png

Opensource LLM

https://github.com/ollama/ollama
https://github.com/facebookresearch/llama

Thinking about AI