linux新M2固态挂载
2024-01-03 10:26:27
一、普通挂载
查看硬盘信息
sudo fdisk -l
创建文件系统
sudo mkfs.ext4 /dev/nvme0n1
创建挂载点
sudo mkdir /home/zain
挂载
sudo mount /dev/nvme0n1 /home/zain
二、永久挂载
vi /etc/fstab
insert:
/dev/nvme0n1 /home/zain ext4 defaults 0 2
wq
sudo mount -a
三、Linux打包
Linux打包整个文件
tar -cvf zain.tar /home/zain
create
filename
-z 用gzip压缩
解压
tar -xzvf zain.tar.gz
四、芯片
ALU
算术逻辑单元(加减乘除、与或非、位运算、移位、条件)
五、Linux文件互传
scp /home/zain.zip root@192.168.6.111:/home/zain
#!/bin/bash
source /usr/local/Ascend/ascend-toolkit/set_env.sh
source /home/ma-user/.bashrc
mkdir llama2-13b
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/config.json
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/generation_config.json
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/pytorch_model.bin.index.json
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/special_tokens_map.json
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/tokenizer_config.json
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/train-00000-of-00001-a09b74b3ef9c3b56.parquet
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/pytorch_model-00001-of-00003.bin
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/pytorch_model-00002-of-00003.bin
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/pytorch_model-00003-of-00003.bin
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/tokenizer.model
wget -P ./llama2-13b https://obs-whaicc-fae-public.obs.cn-central-221.ovaijisuan.com/checkpoint/huggingface/LLaMA/LLaMA2-Chinese-13b-Chat/pretrain_llama2_13B_ptd_8p.sh
python tools/ckpt_convert/llama/convert_weights_from_huggingface.py --input-model-dir /home/ma-user/modelarts/user-job-dir/AscendSpeed/llama2-13b \
--output-model-dir /home/ma-user/modelarts/user-job-dir/AscendSpeed/ckpt \
--tensor-model-parallel-size 1 \
--pipeline-model-parallel-size 1 \
--type 13B \
--deepspeed
mkdir dataset_llama2
cp ./llama2-13b/train-00000-of-00001-a09b74b3ef9c3b56.parquet ./dataset_llama2 -rf
mkdir alpaca_preprocessed
python tools/preprocess_data.py --input /home/ma-user/modelarts/user-job-dir/AscendSpeed/dataset_llama2/train-00000-of-00001-a09b74b3ef9c3b56.parquet \
--output-prefix /home/ma-user/modelarts/user-job-dir/AscendSpeed/alpaca_preprocessed/alpaca \
--tokenizer-type PretrainedFromHF \
--tokenizer-name-or-path llama2-13b \
--tokenizer-not-use-fast \
--handler-name GeneralInstructionHandler
echo "test_ls_1"
echo "test_ls_1"
ls
echo "test_ls_1"
echo "test_ls_1"
echo "test_ls"
echo "test_ls"
ls /home/ma-user/modelarts/user-job-dir/AscendSpeed/alpaca_preprocessed
echo "test_ls"
echo "test_ls"
mv ./llama2-13b/pretrain_llama2_13B_ptd_8p.sh examples/llama2/ -f
bash examples/llama2/pretrain_llama2_13B_ptd_8p.sh
文章来源:https://blog.csdn.net/weixin_44659309/article/details/135343635
本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。 如若内容造成侵权/违法违规/事实不符,请联系我的编程经验分享网邮箱:veading@qq.com进行投诉反馈,一经查实,立即删除!
本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。 如若内容造成侵权/违法违规/事实不符,请联系我的编程经验分享网邮箱:veading@qq.com进行投诉反馈,一经查实,立即删除!