Setup guide for smooth-brain noob dummies from base installation of ubuntu to working inference #2900
Unanswered
iculverr
asked this question in
Community | Q&A
Replies: 6 comments 3 replies
-
installing the docker image |
Beta Was this translation helpful? Give feedback.
0 replies
-
Beta Was this translation helpful? Give feedback.
0 replies
-
Beta Was this translation helpful? Give feedback.
0 replies
-
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Description: I am running on Ubuntu (the latest distro available from Feb 24 2023) with an RTX3090. I want to load up for example GPT-2 or GPT-3 for simple text generation to get things running so that I can then start experimenting and learning. There is no better way to start than to have a working product that you can modify to figure out what the functions do. The current "quick demo" does not provide a clear explanation from beginning to end for beginners on what commands to input and why, where to input them, how to get the output, and what the expected output would look like.
What I have done so far is FIRST git clone the ColossalAI repository and I also installed the ColossalAI with pip install colossalai.
THEN I input colossalai run --nproc_per_node 1 train.py and it points to /home/dev/ and cannot find a train.py, I imagine I need to run the command within the specific directory of resnet or the examples, but nowhere does it say this. Now that I look at the repository it says that it is deprecated and archived. So why does the "quick demo" point to it? How do I quickly demo?
Location: When I test out the "quick demo" from https://colossalai.org/docs/get_started/run_demo and the information at: https://github.com/hpcaitech/ColossalAI-Examples/tree/main/image/resnet
The README.md within /ColossalAI/examples/tutorial tells me to use a conda virtual environment but it does not say this on the main ColossalAI installation page nor on the colossalai.org website
After digging through the files I found the README.md file in /home/dev/ColossalAI/examples/tutorial/auto_parallel
and the README.md file in /home/dev/ColossalAI/examples/tutorial/opt/opt that contains something more like what I am looking for. But WHY is it not on the front page of the github??? Why is it not in the "quick demo" website?
https://github.com/hpcaitech/ColossalAI#Installation
Expectations:
What I am looking for is a step by step full page commands that I can input into the Linux terminal sequentially from base installation to running my first model and getting an output for text generation. The exact model doesn't matter too much, anything GPT-like or BERT-like or BLOOM-like will do.
Areas of future guidance would also be helpful after this first demo works. For example, if I GPT-2 or GPT-3 is successfully loaded and I wanted to branch into making a Chat GPT replica or something with more persistent memory describe conceptually the actions I would need to take in terms of scripting, directory tinkering, and commands of interest.
I expect that the tutorial contains a list of all the installation requirements necessary to load and run the model, the front page says something about NOT install transformers or pytorch at the beginning which is also confusing because it is suggested that these (or the CUDA files) are loaded when colossal ai is loaded??
Screenshots:
image
image
image
image
What is the purpose of the Docker? Do I need it to run inference? It is not mentioned anywhere in the tutorial!
Suggestions:
What I am looking for is a step by step full page commands that I can input into the Linux terminal sequentially from base installation to running my first model and getting an output that is explained like I am five years old and smooth brained.
Beta Was this translation helpful? Give feedback.
All reactions