Script started on 2021-03-10 16:18:12+00:00 [TERM="xterm" TTY="/dev/pts/0" COLUMNS="149" LINES="43"] ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ipmctl show -memoryresources MemoryType | DDR | DCPMM | Total ======================================================== Volatile | 192.000 GiB | 0.000 GiB | 192.000 GiB AppDirect | - | 504.000 GiB | 504.000 GiB Cache | 0.000 GiB | - | 0.000 GiB Inaccessible | - | 1.689 GiB | 1.689 GiB Physical | 192.000 GiB | 505.689 GiB | 697.689 GiB ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl create-namespace -r region0 --mode=devdax { "dev":"namespace0.0", "mode":"devdax", "map":"dev", "size":"248.06 GiB (266.35 GB)", "uuid":"9ad75b6c-52f1-4c58-94e7-a59b937a0a3e", "daxregion":{ "id":0, "size":"248.06 GiB (266.35 GB)", "align":2097152, "devices":[ { "chardev":"dax0.0", "size":"248.06 GiB (266.35 GB)", "target_node":2, "mode":"devdax" } ] }, "align":2097152 } ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl list --regions --namespaces --human --buses { "provider":"ACPI.NFIT", "dev":"ndbus0", "scrub_state":"idle", "regions":[ { "dev":"region1", "size":"252.00 GiB (270.58 GB)", "available_size":"252.00 GiB (270.58 GB)", "max_available_extent":"252.00 GiB (270.58 GB)", "type":"pmem", "iset_id":"0xbcd6eeb86a8f2444", "persistence_domain":"memory_controller" }, { "dev":"region0", "size":"252.00 GiB (270.58 GB)", "available_size":0, "max_available_extent":0, "type":"pmem", "iset_id":"0x1c72eeb8e48b2444", "persistence_domain":"memory_controller", "namespaces":[ { "dev":"namespace0.0", "mode":"devdax", "map":"dev", "size":"248.06 GiB (266.35 GB)", "uuid":"9ad75b6c-52f1-4c58-94e7-a59b937a0a3e", "chardev":"dax0.0", "align":2097152 } ] } ] } ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# daxctl reconfigure-device dax0.0 --mode=system-ram dax0.0: WARNING: detected a race while onlining memory Some memory may not be in the expected zone. It is recommended to disable any other onlining mechanisms, and retry. If onlining is to be left to other agents, use the --no-online option to suppress this warning dax0.0: all memory sections (248) already online [ { "chardev":"dax0.0", "size":266352984064, "target_node":2, "mode":"system-ram", "movable":false } ] reconfigured 1 device ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# mount -t tmpfs -o size=4g,mpol=bind:2 tmpfs /mnt/pmem0 ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl list --regions --namespaces --human --buses { "provider":"ACPI.NFIT", "dev":"ndbus0", "scrub_state":"idle", "regions":[ { "dev":"region1", "size":"252.00 GiB (270.58 GB)", "available_size":"252.00 GiB (270.58 GB)", "max_available_extent":"252.00 GiB (270.58 GB)", "type":"pmem", "iset_id":"0xbcd6eeb86a8f2444", "persistence_domain":"memory_controller" }, { "dev":"region0", "size":"252.00 GiB (270.58 GB)", "available_size":0, "max_available_extent":0, "type":"pmem", "iset_id":"0x1c72eeb8e48b2444", "persistence_domain":"memory_controller", "namespaces":[ { "dev":"namespace0.0", "mode":"devdax", "map":"dev", "size":"248.06 GiB (266.35 GB)", "uuid":"9ad75b6c-52f1-4c58-94e7-a59b937a0a3e", "chardev":"dax0.0", "align":2097152 } ] } ] } ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl disable-namespace namespace0.0 disabled 1 namespace ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl destroy-namespace -f namespace0.0 destroyed 1 namespace ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl list -NR [ { "dev":"region1", "size":270582939648, "available_size":270582939648, "max_available_extent":270582939648, "type":"pmem", "iset_id":-4839418273733860284, "persistence_domain":"memory_controller" }, { "dev":"region0", "size":270582939648, "available_size":270582939648, "max_available_extent":270582939648, "type":"pmem", "iset_id":2049963258282714180, "persistence_domain":"memory_controller" } ] ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl list --regions --namespaces --human --buses { "provider":"ACPI.NFIT", "dev":"ndbus0", "scrub_state":"idle", "regions":[ { "dev":"region1", "size":"252.00 GiB (270.58 GB)", "available_size":"252.00 GiB (270.58 GB)", "max_available_extent":"252.00 GiB (270.58 GB)", "type":"pmem", "iset_id":"0xbcd6eeb86a8f2444", "persistence_domain":"memory_controller" }, { "dev":"region0", "size":"252.00 GiB (270.58 GB)", "available_size":"252.00 GiB (270.58 GB)", "max_available_extent":"252.00 GiB (270.58 GB)", "type":"pmem", "iset_id":"0x1c72eeb8e48b2444", "persistence_domain":"memory_controller" } ] } ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ndctl create-namespace -r region0 --mode=devdax libndctl: ndctl_dax_enable: dax0.1: failed to enable Error: namespace0.0: failed to enable failed to create namespace: No such device or address ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# pwd /root ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# ll ndctl* -rw-r--r-- 1 root root 4096 Mar 10 16:19 ndctl_cmd_log.txt ndctl: total 2120 drwxr-xr-x 16 root root 4096 Dec 24 16:02 ./ drwx------ 20 root root 4096 Mar 10 16:18 ../ -rw-r--r-- 1 root root 52286 Dec 24 15:58 aclocal.m4 -rwxr-xr-x 1 root root 790 Dec 24 15:32 autogen.sh* drwxr-xr-x 2 root root 4096 Dec 24 15:58 autom4te.cache/ drwxr-xr-x 2 root root 4096 Dec 24 15:51 build-aux/ drwxr-xr-x 11 root root 4096 Dec 24 15:32 ccan/ -rw-r--r-- 1 root root 4952 Dec 24 16:01 config.h -rw-r--r-- 1 root root 4602 Dec 24 15:58 config.h.in -rw-r--r-- 1 root root 4549 Dec 24 15:55 config.h.in~ -rw-r--r-- 1 root root 40705 Dec 24 16:01 config.log -rwxr-xr-x 1 root root 60575 Dec 24 16:01 config.status* -rwxr-xr-x 1 root root 498738 Dec 24 15:58 configure* -rw-r--r-- 1 root root 8196 Dec 24 15:32 configure.ac -rwxr-xr-x 1 root root 473354 Dec 24 15:58 configure.lineno* drwxr-xr-x 2 root root 4096 Dec 24 15:32 contrib/ -rw-r--r-- 1 root root 2275 Dec 24 15:32 CONTRIBUTING.md -rw-r--r-- 1 root root 26517 Dec 24 15:32 COPYING drwxr-xr-x 5 root root 4096 Dec 24 16:02 daxctl/ drwxr-xr-x 4 root root 4096 Dec 24 15:32 Documentation/ drwxr-xr-x 8 root root 4096 Dec 24 16:02 .git/ -rw-r--r-- 1 root root 932 Dec 24 15:32 .gitignore -rwxr-xr-x 1 root root 935 Dec 24 15:32 git-version* -rwxr-xr-x 1 root root 290 Dec 24 15:32 git-version-gen* -rw-r--r-- 1 root root 24508 Dec 24 16:02 libccan.a -rwxr-xr-x 1 root root 342320 Dec 24 16:01 libtool* -rw-r--r-- 1 root root 378338 Dec 24 16:02 libutil.a drwxr-xr-x 2 root root 4096 Dec 24 15:32 licenses/ drwxr-xr-x 2 root root 4096 Dec 24 15:50 m4/ -rw-r--r-- 1 root root 42953 Dec 24 16:01 Makefile -rw-r--r-- 1 root root 2374 Dec 24 15:32 Makefile.am -rw-r--r-- 1 root root 919 Dec 24 15:32 Makefile.am.in -rw-r--r-- 1 root root 43990 Dec 24 15:58 Makefile.in -rwxr-xr-x 1 root root 717 Dec 24 15:32 make-git-snapshot.sh* drwxr-xr-x 6 root root 4096 Dec 24 16:02 ndctl/ -rw-r--r-- 1 root root 4214 Dec 24 15:32 ndctl.spec.in -rw-r--r-- 1 root root 1619 Dec 24 15:32 nfit.h -rw-r--r-- 1 root root 3764 Dec 24 15:32 README.md drwxr-xr-x 2 root root 4096 Dec 24 15:32 rhel/ -rwxr-xr-x 1 root root 196 Dec 24 15:32 rpmbuild.sh* drwxr-xr-x 2 root root 4096 Dec 24 15:32 sles/ -rw-r--r-- 1 root root 23 Dec 24 16:01 stamp-h1 drwxr-xr-x 3 root root 4096 Dec 24 16:01 test/ -rw-r--r-- 1 root root 2470 Dec 24 15:32 test.h -rw-r--r-- 1 root root 626 Dec 24 15:32 .travis.yml drwxr-xr-x 4 root root 4096 Dec 24 16:02 util/ -rw-r--r-- 1 root root 31 Dec 24 15:55 version.m4 ]0;root@cloud-dev-benz: ~root@cloud-dev-benz:~# history 1014 kubectl get pods 1015 kubectl get pods|wc -l 1016 kubectl get pods 1017 numactl -H 1018 kubectl apply -f pytorch-kata.yaml 1019 kubectl get pods 1020 numactl -H 1021 kubectl get pods 1022 numactl -H 1023 kubectl get pods 1024 kubectl get pods|wc -l 1025 kubectl get pods 1026 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 7 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1027 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 10 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1028 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 5 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1029 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 2 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1030 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 1 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1031 kubectl get pods 1032 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -q 1 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1033 curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1034 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1035 kubectl get pods 1036 numactl -H 1037 vim pytorch-kata.yaml 1038 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1039 kubectl get pods 1040 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1041 env |grep proxy 1042 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1043 top 1044 kubectl get pods 1045 kubectl delete -f pytorch-kata.yaml 1046 kubectl get pods 1047 ps -ef|grep cloud 1048 kubectl get pods 1049 ps -ef|grep cloud 1050 kubectl get pods 1051 kubectl get ksvc 1052 kubectl get svc 1053 kubectl get pods 1054 systemctl restart kubelet 1055 kubectl get pods 1056 ps -ef|grep cloud 1057 kubectl apply -f pytorch-kata.yaml 1058 kubectl get pods 1059 numactl -H 1060 kubectl get pods 1061 numactl -H 1062 ps -ef|grep cloud 1063 numactl -H 1064 kubectl get pods 1065 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1066 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 3 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1067 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 50 -q 3 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1068 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 50 -q 4 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1069 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -q 20 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1070 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 200 -q 1 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1071 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 300 -q 1 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1072 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 400 -q 1 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1073 kubectl delete -f pytorch-kata.yaml 1074 vim pytorch-kata.yaml 1075 kubectl get pods 1076 ps -ef|grep cloud 1077 vim /opt/kata/bin/cloud-hypervisor 1078 kubectl apply -f pytorch-kata.yaml 1079 numactl -H 1080 kubectl get pods 1081 kubectl get pods|wc -l 1082 kubectl get pods 1083 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1084 kubectl delete -f pytorch-kata.yaml 1085 vim /opt/kata/bin/cloud-hypervisor 1086 numactl -H 1087 kubectl get pods 1088 ps -ef|grep cloud 1089 kubectl get pods 1090 kubectl apply -f pytorch-kata.yaml 1091 kubectl get pods 1092 numactl -H 1093 kubectl get pods 1094 numactl -H 1095 kubectl get pods 1096 numactl -H 1097 kubectl get pods 1098 numactl -H 1099 kubectl get pods 1100 numactl -H 1101 kubectl get pods 1102 numactl -H 1103 kubectl get pods 1104 numactl -H 1105 kubectl get pods 1106 numactl -H 1107 kubectl get pods 1108 numactl -H 1109 kubectl get pods 1110 numactl -H 1111 kubectl get pods 1112 numactl -H 1113 kubectl get pods 1114 numactl -H 1115 kubectl get pods 1116 numactl -H 1117 kubectl get pods 1118 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1119 kubectl delete -f pytorch-kata.yaml 1120 kubectl get pods 1121 vim /opt/kata/bin/cloud-hypervisor 1122 kubectl get pods 1123 kubectl apply -f pytorch-kata.yaml 1124 vim pytorch-kata.yaml 1125 kubectl delete -f pytorch-kata.yaml 1126 vim pytorch-kata.yaml 1127 kubectl get pods 1128 kubectl get pods|wc -l 1129 vim pytorch-kata.yaml 1130 kubectl apply -f pytorch-kata.yaml 1131 kubectl get pods 1132 numactl -H 1133 ps -ef|grep cloud 1134 numactl -H 1135 kubectl delete -f pytorch-kata.yaml 1136 numactl -H 1137 kubectl get pods 1138 numactl -H 1139 kubectl get pods 1140 ps -ef|grep cloud 1141 pkill -9 -f cloud-hypervisor 1142 ps -ef|grep cloud 1143 numactl -H 1144 vim pytorch-kata.yaml 1145 kubectl apply -f pytorch-kata.yaml 1146 vim pytorch-kata.yaml 1147 kubectl get pods 1148 numactl -H 1149 kubectl get pods 1150 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1151 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 220 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1152 kubectl delete -f pytorch-kata.yaml 1153 vim /opt/kata/bin/cloud-hypervisor 1154 kubectl get pods 1155 vim pytorch-kata.yaml 1156 kubectl get pods 1157 kubectl get pods|wc -l 1158 kubectl get pods 1159 ps -ef|grep cloud 1160 cat /opt/kata/bin/cloud-hypervisor 1161 kubectl apply -f pytorch-kata.yaml 1162 numactl -H 1163 ps -ef|grep cloud 1164 numactl -H 1165 kubectl get pods 1166 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 110 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1167 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 66 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1168 kubectl delete -f pytorch-kata.yaml 1169 vim /etc/containerd/config.toml 1170 kubectl get pods 1171 systemctl restart containerd 1172 ls 1173 cd .. 1174 ls 1175 vim inferenceservice.yaml 1176 kubectl apply -f inferenceservice.yaml 1177 cd - 1178 ls 1179 vim pytorch-kata.yaml 1180 kubectl apply -f pytorch-kata.yaml 1181 kubectl get pods 1182 kubectl edit ksvc 1183 kubectl get pods 1184 kubectl get ksvc 1185 kubectl get pods 1186 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1187 kubectl get pods 1188 kubectl get ksvc 1189 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1190 kubectl get ksvc 1191 vim ../inferenceservice.yaml 1192 kubectl create secret `container-registry` --docker-server=https://docker.io/ --docker-email=375686234@qq.com --docker-username=arronwang --docker-password=arron12345 1193 kubectl create secret docker-registry myRegistry --docker-server=https://docker.io/ --docker-email=375686234@qq.com --docker-username=arronwang --docker-password=arron12345 1194 kubectl create secret docker-registry arron.wang --docker-server=https://docker.io/ --docker-email=375686234@qq.com --docker-username=arronwang --docker-password=arron12345 1195 kubectl get secret [arron.wang] --output=yaml 1196 kubectl get secret [REGISTRY-CRED-SECRETS] --output=yaml 1197 kubectl get secret --output=yaml 1198 kubectl get secret [docker-registry] --output=yaml 1199 kubectl create secret docker-registry --docker-server=https://docker.io/ --docker-email=375686234@qq.com --docker-username=arronwang --docker-password=arron12345 1200 kubectl delete -f pytorch-kata.yaml 1201 kubectl get pods 1202 kubectl get ksvc 1203 kubectl apply -f pytorch-kata.yaml 1204 kubectl describe ksvc 1205 kubectl apply -f pytorch-kata.yaml 1206 kubectl describe ksvc 1207 kubectl delete -f pytorch-kata.yaml 1208 kubectl apply -f pytorch-kata.yaml 1209 kubectl describe ksvc 1210 curl https://index.docker.io/v2/arronwang/pytorchserver/manifests/latest 1211 env|grep proxy 1212 export https_proxy=http:// 1213 export https_proxy=http://child-prc.intel.com:913/ 1214 curl https://index.docker.io/v2/arronwang/pytorchserver/manifests/latest 1215 `container-registry` 1216 kubectl create secret docker-registry container-registry --docker-server=https://docker.io/ --docker-email=375686234@qq.com --docker-username=arronwang --docker-password=arron12345 1217 kubectl get secret [container-registry] --output=yaml 1218 kubectl get secret container-registry --output=yaml 1219 kubectl patch serviceaccount default -p "{\"imagePullSecrets\": [{\"name\": \"container-registry\"}]}" 1220 kubectl get pods 1221 kubectl get ksvc 1222 kubectl delete -f pytorch-kata.yaml 1223 kubectl get ksvc 1224 kubectl describe ksvc 1225 kubectl apply -f pytorch-kata.yaml 1226 kubectl describe ksvc 1227 kubectl delete -f pytorch-kata.yaml 1228 kubectl apply -f pytorch-kata.yaml 1229 kubectl describe ksvc 1230 kubectl get ksvc 1231 kubectl get pods 1232 kubectl descibe pod 1233 kubectl describe pod 1234 kubectl describe deployment 1235 vim pytorch-kata.yaml 1236 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1237 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1238 env|grep proxy 1239 unset https_proxy 1240 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1241 kubectl get ksvc 1242 vim /etc/hosts 1243 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1244 time curl -v -d @./input.json http://pytorch-cifar10-predictor-default.default.example.com/v1/models/pytorch-cifar10:predict 1245 ls 1246 vim pytorch.yaml 1247 vim pytorch-kata.yaml 1248 kubectl describe route 1249 time curl -v -d @./input.json http://pytorch-cifar10-predictor-default.default.svc.cluster.local/v1/models/pytorch-cifar10:predict 1250 kubectl get pods 1251 kubectl get ksvc 1252 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 100 -q 1 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1253 time curl -v -d @./input.json http://pytorch-cifar10-predictor-default.default.svc.cluster.local/v1/models/pytorch-cifar10:predict 1254 time curl -v -d @./input.json http://pytorch-cifar10-predictor-default.default.example.com/v1/models/pytorch-cifar10:predict 1255 ls 1256 vim pytorch.yaml 1257 vim pytorch-kata.yaml 1258 time curl -v -d @./input.json http://pytorch-cifar10-predictor-default.default.example.com/v1/models/pytorch-cifar10:predict 1259 vim pytorch-kata.yaml 1260 time curl -v -d @./input.json http://pytorch-cifar10-predictor-default.default.example.com/v1/models/pytorch-cifar10:predict 1261 kubectl get svc istio-ingressgateway --namespace istio-system 1262 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1263 kubectl get pods 1264 ps -ef|grep python 1265 ps -ef|grep pytorchserver 1266 ps -ef|grep pytorchserver|wc -l 1267 ls 1268 kubectl delete -f pytorch-kata.yaml 1269 vim pytorch-kata.yaml 1270 kubectl get pods 1271 vim pytorch-kata.yaml 1272 kubectl get pods 1273 vim /etc/containerd/config.toml 1274 ps -ef|grep runc 1275 kubectl get pods 1276 ps -ef|grep python 1277 kubectl get pods 1278 kubectl describe pod 1279 kubectl get pods 1280 vim /etc/containerd/config.toml 1281 reboot 1282 modprobe br_netfilter 1283 lsmod 1284 kubectl get pods 1285 kubectl get pods --all-namespace 1286 kubectl get pods --all-namespaces 1287 htop 1288 cd knative/kfserving/ 1289 vim docs/samples/v1alpha2/pytorch/README.md 1290 cd docs/samples/v1alpha2/pytorch/ 1291 ls 1292 vim pytorch.yaml 1293 vim perf.yaml 1294 kubectl apply -f perf.yaml 1295 vim README.md 1296 kubectl create -f perf.yaml 1297 kubectl get pods 1298 ls 1299 vim perf.yaml 1300 kubectl get pods 1301 kubectl describe pod pytorch-load-test2hq8l-ndv7l 1302 kubectl get pods 1303 kubectl logs pytorch-load-test2hq8l-ndv7l 1304 vim README.md 1305 kubectl logs pytorch-load-test2hq8l-ndv7l 1306 kubectl delete -f perf.yaml 1307 kubectl logs pytorch-load-test2hq8l-ndv7l 1308 kubectl delete -f perf.yaml 1309 nslookup pytorch-cifar10.default.svc.cluster.local 1310 kubectl describe route 1311 kubectl logs pytorch-load-test2hq8l-ndv7l 1312 kubectl describe route 1313 kubectl describe ksvc 1314 ls 1315 vim pytorch.yaml 1316 vim perf.yaml 1317 kubectl create -f perf.yaml 1318 kubectl get pods 1319 kubectl logs pytorch-load-test5b7gc-gsh9g 1320 kubectl delete -f perf.yaml 1321 ls 1322 vim pytorch.yaml 1323 ls 1324 vim input.json 1325 vim README.md 1326 htop 1327 cd ~/knative/kfserving/pytorch 1328 kubectl get ksvc 1329 vim pytorch-kata.yaml 1330 kubectl apply -f pytorch-kata.yaml 1331 kubectl describe ksvc 1332 kubectl get pods 1333 kubectl describe ksvc 1334 kubectl get pods 1335 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1336 kubectl get svc istio-ingressgateway --namespace istio-system 1337 unset http_proxy 1338 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1339 htop 1340 ls 1341 vim pytorch-kata.yaml 1342 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1343 kubectl delete -f pytorch-kata.yaml 1344 vim pytorch-kata.yaml 1345 kubectl get deployment 1346 kubectl get pods 1347 kubectl delete pod pytorch-load-test2hq8l-ndv7l pytorch-load-test5b7gc-gsh9g 1348 kubectl get pods 1349 kubectl describe pods 1350 kubectl get pods 1351 cp pytorch-kata.yaml pytorch-runc.yaml 1352 vim pytorch-runc.yaml 1353 kubectl get pods 1354 kubectl apply -f pytorch-runc.yaml 1355 kubectl get pods 1356 time curl -v -H "Host: pytorch-cifar10-predictor-default.default.example.com" -d @./input.json http://192.160.172.178/v1/models/pytorch-cifar10:predict 1357 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 10 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1358 kubectl get pods 1359 kubectl delete -f pytorch-runc.yaml 1360 kubectl get pods 1361 vim /etc/containerd/config.toml 1362 kubectl get pods 1363 kubectl delete pod pytorch-cifar10-predictor-default-6vj6s-deployment-7cd84d67srdp 1364 kubectl get pods 1365 ps -ef|grep python 1366 kill 45251 1367 ps -ef|grep python 1368 kill -9 45251 1369 ps -ef|grep python 1370 kubectl get pods 1371 vim pytorch-kata.yaml 1372 vim /opt/kata/bin/cloud-hypervisor 1373 vim pytorch-kata.yaml 1374 kubectl apply -f pytorch-kata.yaml 1375 kubectl get pods 1376 ps -ef|grep cloud 1377 kubectl get pods 1378 ps -ef|grep cloud 1379 kubectl get pods 1380 ps -ef|grep cloud 1381 ps -ef|grep python 1382 ps -ef|grep cloud 1383 kubectl delete -f pytorch-kata.yaml 1384 kubectl get pods 1385 ps -ef|grep python 1386 kubectl get pods 1387 numactl -H 1388 kubectl get pods 1389 vim /etc/containerd/config.toml 1390 systemctl restart containerd 1391 vim pytorch-kata.yaml 1392 kubectl apply -f pytorch-kata.yaml 1393 cat /opt/kata/bin/cloud-hypervisor 1394 kubectl get pods 1395 kubectl get ksvc 1396 kubectl get pods 1397 numactl -H 1398 daxctl reconfigure-device --mode=system-ram dax0.0 1399 cd /sys/fs/cgroup 1400 numactl -H 1401 echo "0-2" > cpuset/kubepods/cpuset.mems 1402 echo "0-2" > cpuset/kubepods/besteffort/cpuset.mems 1403 echo "0-2" > cpuset/kubepods/burstable/cpuset.mems 1404 kubectl get pods 1405 ps -ef|grep cloud 1406 kubectl get pods 1407 numactl -H 1408 kubectl get pods 1409 numactl -H 1410 kubectl get pods 1411 cd - 1412 kubectl delete -f pytorch-kata.yaml 1413 kubectl get pods 1414 ps -ef|grep cloud 1415 pkill -9 -f cloud-hypervisor 1416 ps -ef|grep cloud 1417 ps -ef|grep hyper 1418 ps -ef|grep hyper|wc -l 1419 kubectl get pods 1420 kubectl get pods|wc -l 1421 kubectl get pods 1422 ps -ef|grep virtio 1423 ps -ef 1424 ps -ef|grep containerd-shim-kata-v2 1425 ps -ef|grep containerd-shim-kata-v2|wc -l 1426 kubectl get pods 1427 pkill -9 -f containerd-shim-kata-v2 1428 kubectl get pods 1429 ps -ef|grep cloud 1430 ps -ef|grep virtiof 1431 kubectl get pods 1432 kubectl apply -f pytorch-kata.yaml 1433 kubectl get pods 1434 numactl -H 1435 kubectl get pods 1436 kubectl describe pod pytorch-cifar10-predictor-default-jvl9z-deployment-6b99bd4prlrs 1437 ps -ef|grep cloud 1438 kubectl describe pod pytorch-cifar10-predictor-default-jvl9z-deployment-6b99bd4prlrs 1439 cd /sys/fs/cgroup 1440 cat cpuset/kubepods/cpuset.mems 1441 cat cpuset/kubepods/besteffort/cpuset.mems 1442 cat cpuset/kubepods/burstable/cpuset.mems 1443 kubectl describe pod pytorch-cifar10-predictor-default-jvl9z-deployment-6b99bd4prlrs 1444 kubectl get pods 1445 cd - 1446 ls 1447 kubectl delete -f pytorch-kata.yaml 1448 ps -ef|grep cloud 1449 kubectl get pods 1450 ps -ef|grep cloud 1451 vim pytorch-kata.yaml 1452 ps -ef|grep cloud 1453 vim pytorch-kata.yaml 1454 ps -ef|grep cloud 1455 kubectl get pods 1456 pkill -9 -f cloud-hypervisor 1457 kubectl get pods 1458 ps -ef|grep cloud 1459 ps -ef|grep kata 1460 pkill -9 -f containerd-shim-kata-v2 1461 ps -ef|grep kata 1462 kubectl get pods 1463 ps -ef|grep kata 1464 kubectl get pods 1465 vim pytorch-kata.yaml 1466 cat /opt/kata/bin/cloud-hypervisor 1467 numactl -H 1468 ps -ef|grep cloud 1469 kubectl apply -f pytorch-kata.yaml 1470 kubectl get pods 1471 kubectl describe pod pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1472 kubectl get pods 1473 kubectl describe pod pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1474 kubectl get pods 1475 kubectl describe pod pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1476 kubectl get pods 1477 kubectl describe pod pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1478 kubectl get pods 1479 kubectl describe pod pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1480 kubectl get pods 1481 kubectl logs pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1482 kubectl logs pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 kfserving-container 1483 kubectl logs pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 queue-proxy 1484 kubectl delete -f pytorch-kata.yaml 1485 cd ../ 1486 find . -name inferenceservice.yaml 1487 kubectl apply -f ./config/configmap/inferenceservice.yaml 1488 cd - 1489 vim pytorch-kata.yaml 1490 kubectl apply -f pytorch-kata.yaml 1491 kubectl get pods 1492 kubectl describe pod pytorch-cifar10-predictor-default-25nzv-deployment-86d54956rcnn 1493 kubectl get pods 1494 numactl -H 1495 kubectl get pods 1496 numactl -H 1497 kubectl get pods 1498 numactl -H 1499 kubectl get pods 1500 kubectl delete pod pytorch-cifar10-predictor-default-bt8tp-deployment-76c4f49zmc55 1501 kubectl get pods 1502 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 22 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1503 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 33 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1504 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 44 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1505 kubectl get pods 1506 kubectl delete -f pytorch-kata.yaml 1507 kubectl get pods 1508 ps -ef|grep cloud 1509 pkill -9 -f cloud-hypervisor 1510 ps -ef|grep kata 1511 kill -9 193326 1512 ps -ef|grep kata 1513 ps -ef|grep cloud 1514 kubectl get pods 1515 kubectl apply -f pytorch-kata.yaml 1516 cat pytorch-kata.yaml 1517 kubectl get pods 1518 numactl -H 1519 kubectl get pods 1520 numactl -H 1521 kubectl get pods 1522 numactl -H 1523 kubectl get pods 1524 numactl -H 1525 kubectl get pods 1526 numactl -H 1527 kubectl get pods 1528 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 44 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1529 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 50 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1530 kubectl delete -f pytorch-kata.yaml 1531 vim pytorch-kata.yaml 1532 kubectl get pods 1533 ps -ef|grep cloud 1534 pkill -9 -f cloud-hypervisor 1535 ps -ef|grep cloud 1536 ps -ef|grep kata 1537 kubectl apply -f pytorch-kata.yaml 1538 cat pytorch- 1539 cat pytorch-kata.yaml 1540 kubectl get pods 1541 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 30 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1542 numactl --cpunodebind=1 --membind=1 ./hey -cpus 10 -z 30s -c 40 -m POST -D ./input.json -host "pytorch-cifar10-predictor-default.default.example.com" "http://192.160.172.178/v1/models/pytorch-cifar10:predict" 1543 kubectl delete -f pytorch-kata.yaml 1544 cd ../ 1545 ls 1546 vim sklearn-full.yaml 1547 htop 1548 kubectl get pods 1549 htop 1550 cd knative/kfserving/ 1551 ls 1552 vim sklearn-full.yaml 1553 vim pytorch/pytorch-kata.yaml 1554 vim sklearn-full.yaml 1555 vim /opt/kata/bin/cloud-hypervisor 1556 kubectl apply -f sklearn-full.yaml 1557 ps -ef|grep cloud 1558 kubectl get pods 1559 kubectl get ksvc 1560 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 100 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1561 numactl -H 1562 ps -ef|grep cloud 1563 cat /opt/kata/bin/cloud-hypervisor 1564 numactl -H 1565 kubectl delete -f sklearn-full.yaml 1566 ps -ef|grep cloud 1567 numactl -H 1568 ps -ef|grep cloud 1569 numactl -H 1570 ps -ef|grep cloud 1571 numactl -H 1572 ps -ef|grep cloud 1573 numactl -H 1574 ps -ef|grep cloud 1575 kubectl get pods 1576 ps -ef|grep cloud 1577 ps -ef|grep kata 1578 numactl -H 1579 vim /opt/kata/bin/cloud-hypervisor 1580 kubectl apply -f sklearn-full.yaml 1581 kubectl get pods 1582 numactl -H 1583 kubectl get pods 1584 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 100 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1585 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 90 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1586 kubectl delete -f sklearn-full.yaml 1587 vim /opt/kata/bin/cloud-hypervisor 1588 vim sklearn-full.yaml 1589 kubectl get pods 1590 ps -ef|grep kata 1591 ps -ef|grep cloud 1592 cat sklearn-full.yaml 1593 kubectl apply -f sklearn-full.yaml 1594 kubectl get pods 1595 numactl -H 1596 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 200 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1597 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 300 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1598 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 500 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1599 kubectl delete -f sklearn-full.yaml 1600 vim /opt/kata/bin/cloud-hypervisor 1601 kubectl get pods 1602 ps -ef|grep cloud 1603 ps -ef|grep kata 1604 numactl -H 1605 kubectl apply -f sklearn-full.yaml 1606 kubectl get pods 1607 kubectl get pods|wc -l 1608 numactl -H 1609 kubectl get pods|wc -l 1610 kubectl get pods 1611 numactl -H 1612 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 500 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1613 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 400 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1614 kubectl get pods 1615 numactl -H 1616 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 250 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1617 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 200 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1618 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 180 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1619 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 160 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1620 kubectl get pods 1621 kubectl get pods|wc -l 1622 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 160 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1623 kubectl get pods 1624 kubectl get pods|wc -l 1625 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 160 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1626 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 100 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1627 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 150 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1628 numactl --cpunodebind=1 --membind=1 ./hey -z 30s -c 120 -m POST -D ./iris-input.json -host "sklearn-iris-predictor-default.default.example.com" "http://192.160.172.178/v1/models/sklearn-iris:predict" 1629 kubectl get pods 1630 cd knative/kfserving/ 1631 kubectl delete -f sklearn-full.yaml 1632 vim /etc/containerd/config.toml 1633 kubectl get pods 1634 vim /etc/containerd/config.toml 1635 systemctl restart containerd 1636 htop 1637 ls 1638 exit 1639 ls 1640 kubectl get pods 1641 ps -ax 1642 ps -aux 1643 exit 1644 htop 1645 ls 1646 cd mp4/ 1647 ls 1648 ls -alh *.mp4 1649 history |grep ffmpeg 1650 numactl -H 1651 rm /tmp/destination.flv 1652 time numactl --cpunodebind=0 --membind=0 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 16 /tmp/destination.flv 1653 rm /tmp/destination.flv 1654 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 16 /tmp/destination.flv 1655 rm /tmp/destination.flv 1656 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 8 /tmp/destination.flv 1657 rm /tmp/destination.flv 1658 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 8 /tmp/destination.flv 1659 rm /tmp/destination.flv 1660 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 4 /tmp/destination.flv 1661 rm /tmp/destination.flv 1662 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 4 /tmp/destination.flv 1663 rm /tmp/destination.flv 1664 time numactl --cpunodebind=0 --membind=0 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 22 /tmp/destination.flv 1665 rm /tmp/destination.flv 1666 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 22 /tmp/destination.flv 1667 ls 1668 rm /tmp/destination.flv 1669 time numactl --cpunodebind=0 --membind=0 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 4 /tmp/destination.flv 1670 rm /tmp/destination.flv 1671 time numactl --cpunodebind=0 --membind=0 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 8 /tmp/destination.flv 1672 ls -alh test_1080p.mp4 1673 rm /tmp/destination.flv 1674 time numactl --cpunodebind=0 --membind=0 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 4 /tmp/destination.flv 1675 rm /tmp/destination.flv 1676 time numactl --cpunodebind=0 --membind=0 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 2 /tmp/destination.flv 1677 rm /tmp/destination.flv 1678 time numactl --cpunodebind=0 --membind=2 ffmpeg -i test_1080p.mp4 -c:v libx264 -crf 30 -threads 2 /tmp/destination.flv 1679 ls 1680 cd knative/ 1681 ls 1682 cd knative-routing-tutorial/ 1683 ls 1684 vim nginx1.yaml 1685 ls 1686 vim nginx_test.yaml 1687 vim helloworld1.yaml 1688 vim test.yaml 1689 vim nginx1.yaml 1690 proxychains 1691 apt install proxychains4 1692 lscpu 1693 uname -a 1694 vim /etc/proxychains4.conf 1695 proxychains ssh -p 12222 arron@139.227.46.229 1696 ls 1697 free -m 1698 w 1699 ps -ef 1700 ps -ef|grep ssh 1701 kill 195572 1702 ps -ef|grep ssh 1703 ls 1704 w 1705 dmidecode -t memory 1706 exit 1707 top 1708 cd /root/wrk/ 1709 ls 1710 make 1711 ls 1712 cp wrk /usr/bin/ 1713 taskset -c 0-25 wrk --latency --timeout 2s -t 100 -c 10000 -d 60s http://10.239.158/82:30001 1714 taskset -c 0-25 wrk --latency --timeout 2s -t 100 -c 10000 -d 60s http://10.239.158.82:30001 1715 ps -aux | grep taskset 1716 ps -aux | grep wrk 1717 kill -9 1597329 1718 taskset -c 0-25 wrk --latency --timeout 2s -t 100 -c 10000 -d 60s http://10.239.85.182:30001 1719 cat /proc/cpuinfo 1720 taskset -c 0-25 wrk --latency --timeout 2s -t 50 -c 10000 -d 60s http://10.239.85.182:30001 1721 lscpu 1722 taskset -c 0-21 wrk --latency --timeout 2s -t 88 -c 10000 -d 60s http://10.239.85.182:30001 1723 taskset -c 0-21 wrk --latency --timeout 2s -t 44 -c 10000 -d 60s http://10.239.85.182:30001 1724 ping 10.239.85.182 1725 taskset -c 0-21 wrk --latency --timeout 2s -t 40 -c 8000 -d 60s http://10.239.85.182:30001 1726 taskset -c 0-21 wrk --latency --timeout 2s -t 10 -c 1000 -d 60s http://10.239.85.182:30001 1727 taskset -c 0-21 wrk --latency --timeout 2s -t 20 -c 2000 -d 60s http://10.239.85.182:30001 1728 uname -r 1729 taskset -c 0-50 wrk --latency --timeout 2s -t 20 -c 2000 -d 60s http://10.239.85.182:30001 1730 ulimit -a 1731 ulimit -n 65535 1732 taskset -c 0-21 wrk --latency --timeout 2s -t 20 -c 2000 -d 60s http://10.239.85.182:30001 1733 taskset -c 0-25 wrk --latency --timeout 2s -t 100 -c 10000 -d 30s http://10.239.85.182:30001 1734 ping 10.239.58.182 1735 ping 10.239.85.182 1736 taskset -c 0-25 wrk --latency --timeout 2s -t 100 -c 10000 -d 30s http://10.239.85.182:30001 1737 ifconfig ` 1738 ifconfig 1739 exit 1740 ls 1741 mkdir tdx 1742 cd tdx/ 1743 git clone 1744 git clone https://gitlab.devtools.intel.com/tdx/linux.git 1745 cd tdx/linux/ 1746 ls 1747 vim KVM-TDX.README.md 1748 pwd 1749 vim .config 1750 make bzImage -j44 1751 sudo apt-get install build-essential libncurses-dev bison flex libssl-dev libelf-dev 1752 cat /etc/lsb-release 1753 apt-get install build-essential libncurses-dev bison flex libssl-dev libelf-dev 1754 make bzImage -j44 1755 cp arch/x86/boot/bzImage . 1756 find -name "*.ko" 1757 scp arch/x86/boot/bzImage tdx@10.239.85.188:/home/tdx/host_kernel/ 1758 ls -alh bzImage 1759 ps -fe|grep -i net 1760 cat /etc/networks 1761 cat /etc/network/interfaces 1762 cat /etc/network/if-up.d/ethtool 1763 ls 1764 vim Makefile 1765 make install 1766 vim /boot/grub/grub.cfg 1767 scp /boot/initrd.img-5.11.0-rc5+ /boot/System.map-5.11.0-rc5+ /boot/vmlinuz-5.11.0-rc5+ tdx@10.239.85.188:/home/tdx/host_kernel/ 1768 make uninstall 1769 efibootmgr 1770 ls /boot/efi/ 1771 vim /boot/grub/grub.cfg 1772 update-initramfs -c -k 5.4.0-65-generic 1773 update-grub 1774 vim /etc/default/grub 1775 ls 1776 ls /boot/ 1777 rm /boot/vmlinuz-5.11.0-rc5+ 1778 rm /boot/initrd.img-5.11.0-rc5+ 1779 rm /boot/System.map-5.11.0-rc5+ 1780 update-grub 1781 vim /boot/grub/grub.cfg 1782 git tag 1783 git checkout tdx-kvm-2021.02.01 1784 ls 1785 scp .config tdx@10.239.85.188:/home/tdx/ 1786 cd tdx/linux/ 1787 ls 1788 vim .config 1789 make bzImage -j64 1790 ls 1791 scp -r root@vt-master:/root/Music . 1792 rmdir --help 1793 sudo apt install cpuid 1794 lscpu 1795 cpuid -1 --leaf=0x15 1796 cpuid -1 --leaf=0x16~ 1797 cpuid -1 --leaf=0x15 1798 exit 1799 cpuid -1 --leaf=0x15 1800 exit 1801 cat /boot/grub/grub.cfg 1802 :q 1803 kubectl get pods 1804 kubectl get pods --all-namespaces 1805 lsmod 1806 ps -ef|grep kubelet 1807 systemctl status kubelet 1808 vim /etc/containerd/config.toml 1809 kubectl get pods 1810 exit 1811 vim /var/lib/kubelet/config.yaml 1812 vim "/usr/lib/systemd/system/kubelet.service.d/10-kubeadm.conf" 1813 vim /usr/lib/systemd/system/kubelet.service.d/10-kubeadm.conf 1814 systemctl status kubelet 1815 vim /etc/systemd/system/kubelet.service.d/10-kubeadm.conf 1816 vim /etc/systemd/system/kubelet.service.d/0-containerd.conf 1817 vim /etc/systemd/system/kubelet.service.d/10-kubeadm.conf 1818 vim ~/config 1819 vim /usr/lib/systemd/system/kubelet.service 1820 vim /etc/systemd/system/kubelet.service.d/10-kubeadm.conf 1821 vim /usr/lib/systemd/system/kubelet.service 1822 vim /etc/systemd/system/kubelet.service.d/10-kubeadm.conf 1823 ls 1824 ms mp4/ 1825 ls mp4/ 1826 cd mp4/ 1827 ls 1828 ls -alh 1829 pwd 1830 ifconfig 1831 pwd 1832 ls 1833 du -h 1834 ls -alh 1835 daxctl 1836 daxctl --list-cmds 1837 daxctl reconfigure-device 1838 daxctl reconfigure-device -m --help 1839 ls 1840 numactl --hardware 1841 uname -a 1842 ipmctl version 1843 ndctl -v 1844 mount 1845 free -m 1846 numactl -H 1847 ls 1848 mount 1849 ls 1850 cd /mnt/ 1851 ls 1852 w 1853 ls 1854 mkdir pmem0 1855 ls 1856 numactl -h 1857 numactl -H 1858 ls 1859 mount 1860 umount pmem0 1861 ls 1862 mount -t tmpfs -o size=4g,mpol=bind:2 tmpfs /mnt/pmem0 1863 ls 1864 mount 1865 numactl --hardware 1866 dd if=/dev/zero of=/mnt/pmem0/A bs=1M count=1024 1867 numactl -H 1868 free -m 1869 rm pmem0/A 1870 free -m 1871 numactl -H 1872 ndctl list 1873 ndctl list -a 1874 ndctl list ls 1875 ndctl list 1876 ndctl destroy-namespace -f namespace0.0 1877 ndctl list 1878 free -m 1879 n