-
Notifications
You must be signed in to change notification settings - Fork 42
Cannot do full build on Centos7 #27
Comments
After trying the commands 3 more times - I get to this point: exit status 1 But running again, once again stops at the Software Order error |
Update No.2 After actually changing the registry to docker ./build.sh --type full --docker-namespace myprivdockerregistryname --docker-registry-url index.docker.io --zip /home/admin/sas-container-recipes-master/SAS_Viya_deployment_data.zip --addons "auth-demo ide-jupyter-python3" I got into the build process (32 builds), and on build fails on: 2019/06/20 20:03:33 Starting 32 build processes ... (this may take several minutes) goroutine 135 [running]: On second attempt: 2019/06/20 20:29:17 Starting 32 build processes ... (this may take several minutes) 2019/06/20 20:29:17 [TIP] System resource utilization can be seen by using the goroutine 130 [running]: Any idea what might be causing this? I am going to try and keep forcing the builds, because I think each time I start a build, the process starts on a different container, so it may be that I can build 31/32 and it will fail on say just one, perhaps isolating the issue? Am I right about it being worth pushing it through? |
Update 3. Another fail this time with panic: runtime error: index out of range goroutine 92 [running]: Is this where it is having issues? Line 688 says while 831 is a loop over order.WorkerCount |
@alanzablocki I'm working on an experimental branch with instructions on how to use ACR+AKS, along with fixes to some of the issues you've encountered. Want to try it out? https://github.com/sassoftware/sas-container-recipes/tree/19m06-dev We could work through the issues from there and I'll keep updating the branch. |
@Collinux Awesome, thanks for letting me know I will try it over the weekend. |
Should I run docker image prune -a, to clean everything out, since I will use a new build branch? |
There's no need to do an image prune since the former images builds do not affect future image builds. If you don't have much disk space then it helps to do an occasional image prune. I use |
Thanks, I actually started a build a few minutes ago to run in the background. Error #1: NAME STATUS LAYER with command [admin@RM-SAS-DOCKER-01 sas-container-recipes]$ ./build.sh --type full --docker-namespace sasfull --docker-registry-url https://sasfull.azurecr.io --zip /home/admin/sas-container-recipes/SAS_Viya_deployment_data.zip --addons "auth-demo ide-jupyter-python3" Note: -n|--docker-namespace|--docker-registry-namespace has been deprecated. This function is no longer performed and its usage can be ignored. ===============================
|
I changed the command and run again: It started to build and then... 2019/06/21 16:09:23 Starting 32 build processes. NAME STATUS LAYER For error details, review the following log files: exit status 1 |
Maybe a network blip? In log.txt
|
Hmm, I will try again. Any chance that I have run out "hits" to my SAS orders. I am not using a mirror url. |
I think it was moving along but then stopped again - I can also see the size of my azure registry increased to 0.24GB, so things are going: Here is the error now: NAME STATUS LAYER For error details, review the following log files: exit status 1 This time I cannot see any obvious errors in the logs. |
I run the same commands again twice and both times I got the registry errors but they are different: exit status 1 b) exit status 1 I run it a third time and so far it is building , so all of the above seem to be passing issues, but they do stop a build. |
Looks like I keep getting the talking to yum issue: here is another NAME STATUS LAYER For error details, review the following log files: exit status 1 I will keep running one error after the other to try and beat this, rather than post more error messages. |
@Collinux It seems that the build stops on a number of SAS base install steps, with repo errors, failed 127.0.0.1 and no more mirrors to try messages. I am attaching the log from a run which connected to sas repositories hosted by sas. We even tried using our http hosted sas_repos folder as a mirror-url to try to exclude connections to sas servers, but we see the same general errors. |
Hello @alanzablocki , this may or may not help but ... in the log.txt, you have "HTTPS Error 410 - Gone". I always recommend to use a local YUM mirror, to avoid this. If you are using a local YUM mirror, I'm curious as to how you serve the HTTP content. In the past, when using a quick-and-dirty mirror served by Python SimpleHTTPServer, I have seen it be overwhelmed by those requests (simplehttpserver is single-threaded). Finally, when using a YUM mirror, beware of hostname resolution. I've been burned before by using " I hope some of this helps. |
example_mirror_error.txt We will try your ip suggestion. How do you "reduce" the number of images being built? |
P.S. @erwangranger Do you know what the maximum number of downloads is? |
I hope this helps. |
aha! Cannot find the file using the browser, it gives a 404. It is the + sign then, and we solved it with, https://helpx.adobe.com/experience-manager/kb/CannotOpenAFileHavingSpecialCharactersInTheFilenameOnIIS.html Now we can download the file via browser, and so will try the build again. |
We hit this yum related error [container.go:431] &{failed: [127.0.0.1] (item=[u'sas-envesntl', u'sas-bootstrap-config', u'sas-runjavasvc']) => {"changed": false, "item": ["sas-envesntl", "sas-bootstrap-config", "sas-runjavasvc"], "msg": "\n\n One of the configured repositories failed (Unknown),\n and yum doesn't have enough cached data to continue. At this point the only\n safe thing yum can do is fail. There are a few ways to work "fix" this:\n\n 1. Contact the upstream for the repository and get them to fix the problem.\n\n 2. Reconfigure the baseurl/etc. for the repository, to point to a working\n upstream. This is most often useful if you are using a newer\n distribution release than is supported by the repository (and the\n packages for the previous distribution release still work).\n\n 3. Run the command with the repository temporarily disabled\n yum --disablerepo= ...\n\n 4. Disable the repository permanently, so yum won't use it by default. Yum\n will then just ignore the repository until you permanently enable it\n again or use --enablerepo for temporary usage:\n\n yum-config-manager --disable \n or\n subscription-manager repos --disable=\n\n 5. Configure the failing repository to be skipped, if it is unavailable.\n Note that yum will try to contact the repo. when it runs most commands,\n so will have to try and fail each time (and thus. yum will be be much\n slower). If it is a very temporary problem though, this is often a nice\n compromise:\n\n yum-config-manager --save --setopt=.skip_if_unavailable=true\n\nCannot find a valid baseurl for repo: base/7/x86_64\n", "rc": 1, "results": []} } I found a few posts that "talk about it" but am unsure how to proceed. Can you advise? We are now using these settings when buidling to avoid some of the error we talked about above; ./build.sh --type full --docker-registry-url sasfull.azurecr.io --mirror-url 'http:///sas_repos/' --zip /home/admin/sas-container-recipes/SAS_Viya_deployment_data.zip --addons "auth-demo ide-jupyter-python3" --workers 1 The error is from builds/full-2019-06-24-22-56-37/sas-viya-dataservices/log.txt Thanks for your help |
@alanzablocki you provided the argument |
@Collinux we had 4 when we started the build , the default , since I did not use the --workers flag initially. We saw this same error with 4. I will try --workers 2. We went down all the way to one to eliminate the possibility that there were too many http requests being placed to our sas_repos mirror url. I am happy with one image at a time. yum clean metadata Bizarrely, I was able to build as many as 3-4 images after doing that, rather than one. I am not convinced that is the solution though. Another day, the build would also keep going for a couple of images, and then stop, with the same errors (cache or talking to yum). Although it seems like maybe network errors, our network has been up all the time. I am at 27 out of 32 images, so getting closer. |
@Collinux Our full build finished sometime over the weekend, 113GB in /data for the 32 docker images (we cut it close as /data had 119GB of space). This takes up 50GB in Azure container registry. Next I will follow your AKS instructions to deploy to Kubernetes. |
@Collinux Did something change in the how we access Jupyter? It is not available on port 8888 anymore, only on 8080, which is strange because usually the kernels die on anything other than 8888, like on 80 or 8080. |
FWIW, I encountered this exact same error for a 27 image build and ultimately solved it by passing "--workers 4". Each time I tried the full build, one of the build's worker processes would inevitably fail - sometimes on the same SAS rpm as the previous attempt, sometimes on a different one. None of the rpm names contained "+" signs. I'm serving my YUM mirror from an nginx Docker container on a separate server, as follows:
Conclusion |
Hi @TylerGillson , thanks for the useful info. Couple things from me:
|
Hi @erwangranger , I tested my hypothesis today using a tuned NGINX container with twice the keepalive_timeout (130 instead of 65) and no --workers argument. build.sh ran to completion in 47m4s. I chose to double the keepalive_timeout arbitrarily, but that seems to have worked! Here's the NGINX Dockerfile I used:
|
Describe the bug
I am unable to do a full build, after successfully doing programming-only single build. I see the message:
2019/06/20 18:54:24 Serving license and entitlement on sas-container-recipes-builder:1976 (172.17.0.2)
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0x889372]
goroutine 23 [running]:
main.(*SoftwareOrder).TestRegistry(0xc0002c8000, 0xc00033c180, 0xc00033c120, 0xc00033c0c0)
/sas-container-recipes/order.go:1046 +0x1e2
created by main.NewSoftwareOrder
/sas-container-recipes/order.go:308 +0x7e3
exit status 2
sas-container-recipes-builder-19.05.0-20190620135121-no-git-sha
Is this complaining about the registry? Can it be azure container registry or must it be an actual dockerhub registry?
Thanks
To Reproduce
Steps to reproduce the behavior:
./build.sh --type full --docker-namespace sasfullxyz --docker-registry-url sasfullxyz.azurecr.io --zip /home/admin/sas-container-recipes-master/SAS_Viya_deployment_data.zip --addons "auth-demo ide-jupyter-python3"
Expected behavior
Get into the build step
Environment (please complete the applicable information):
[admin@RM-SAS-DOCKER-01 sas-container-recipes-master]$ docker version
Client:
Version: 18.09.6
API version: 1.39
Go version: go1.10.8
Git commit: 481bc77156
Built: Sat May 4 02:34:58 2019
OS/Arch: linux/amd64
Experimental: false
Server: Docker Engine - Community
Engine:
Version: 18.09.6
API version: 1.39 (minimum version 1.12)
Go version: go1.10.8
Git commit: 481bc77
Built: Sat May 4 02:02:43 2019
OS/Arch: linux/amd64
Experimental: false
The text was updated successfully, but these errors were encountered: