Running Amazon-Linux Docker Image on MacOS - docker

I was able to download amazon linux's image from this dockerhub -
ambakshi/amazon-linux 2016.03 083f32382cc4 4 months ago 142.7 MB
But I'm not able to do "yum install" any package on it.
Keeps giving a 403 Forbidden error.
[root#8ea0c348a91c /]# yum install lxml
Loaded plugins: priorities, update-motd, upgrade-helper
http://packages.us-west-2.amazonaws.com/2016.09/main/20160901f6a8/x86_64/repodata/repomd.xml?instance_id=fail&region=URLError: [Errno 14] PYCURL ERROR 22 - "The requested URL returned error: 403 Forbidden"
Trying other mirror.
IS there something I'm missing?
I'm trying all of this in Mac OS. Can someone please help?
Thanks
Ramya

Answering my own question:
Amazon has now released official "Amazon Linux" Docker Images which can be used for local development:
Please check the details here:
https://aws.amazon.com/blogs/aws/new-amazon-linux-container-image-for-cloud-and-on-premises-workloads/
http://docs.aws.amazon.com/AmazonECR/latest/userguide/amazon_linux_container_image.html

You can only download those packages if the container is running on Amazon's infrastructure within their network (an EC2 instance for example). It won't work from your local development machine.

Related

How do you resolve the GitLab error "Error response from daemon: invalid condition: 'not-running'"?

I set up a Windows GitLab runner that's supposed to download a Docker image from our Container Registry and then run a build script in the pipeline. Unfortunately the Docker container never launches due to the following error:
Running with gitlab-runner 15.1.0 (76984217)
on WindowsDockerRunner wZMWQZYi
Resolving secrets
Preparing the "docker-windows" executor
Using Docker executor with image mcr.microsoft.com/windows/servercore:ltsc2019 ...
Pulling docker image mcr.microsoft.com/windows/servercore:ltsc2019 ...
Using docker image sha256:e6b07227af5ca9303c2112b574f6f27f38135bbf9df29d829142410221967401 for mcr.microsoft.com/windows/servercore:ltsc2019 with digest mcr.microsoft.com/windows/servercore#sha256:26c6c296a4737ba478fe3c3e531b098f89b5562c40b416ba6fb8177ac462d1af ...
Preparing environment
Running on RUNNER-WZMWQZYI via
runner2...
ERROR: Job failed (system failure): prepare environment: Error response from daemon: invalid condition: "not-running". Check https://docs.gitlab.com/runner/shells/index.html#shell-profile-loading for more information
The error message doesn't clearly state what the cause of the problem is and the documentation that it references doesn't mention anything about "condition". Based on the link pointing to shell profiles I suspect it might have something to do with the shell that's being run, but when I run the Docker container locally it boots into PowerShell just fine.
Does anyone know how to solve this?
I came across this issue after installing Docker Engine using the Windows Server install script, which fetches docker.exe and dockerd.exe from https://master.dockerproject.org, These builds were last updated in March 2022, I found gitlab-runner 14.9 and earlier work okay with this version (released prior to March 2022), but 14.10 does not (released 2022-04-19) nor do any newer versions.
Installing Docker Desktop resolves this as it provides the latest version. However using Docker Desktop introduces licensing issues. An alternative is to manually install Docker Engine / update the version downloaded by the Microsoft script.
Docker Engine builds are provided on the Moby GitHub project to download from https://download.docker.com/win/static/stable/x86_64/ downloading the lastest version from here and replacing the docker executables in C:\Windows\System32 fixes the problem, working with the latest gitlab-runner.
An alternative is to use the docker-engine chocolatey package (which incidentally I maintain) which provides installation scripting for the above stable builds:
choco install docker-engine
There is also an open issue with the Windows-Containers team to move off (out of date) nightlies: https://github.com/microsoft/Windows-Containers/issues/256 which would provide a stable docker build, through the Microsoft recommended installation method.
Was finally able to solve this issue. We had the Docker Engine installed on our GitLab Runner, but that doesn't seem to be sufficient for GitLab CI/CD. After installing Docker Desktop on the runner the issue disappeared and we were able to run the pipeline.
After some trial and error I got it up and running.
I have another server running the gitlab-runner and docker without any issues (no docker desktop installed, which is not allowed because of licensing stuff).
The server I'm trying to setup right now is a 'redundancy' build server.
So to find out what was my problem, I started switching things from one build server to the other. Currently, it appears that simply downgrading to the gitlab-runner V13.4.0 was enough.
I did reregister the runner, since gitlab stated that the V15.x.x version was using executor "unknown".
Not sure what is going on there, but at least I can continue building now.

MISP instance through docker on raspberry pi running Ubuntu 20.04 server

Thanks so much in advance for taking the time to read/provide any advice here.
So, I am trying to get an instance of MISP running through docker. The hardware I have it running on is a raspberry pi 4 running Ubuntu 20.04 (server edition).
I thought I installed all software dependencies, but being new to using docker, perhaps I haven't. I'm using this repository for the docker image: https://github.com/MISP/misp-docker
After running the command sudo docker-compose up after copying the .env file to the root directory, I get the error that I am going to post an image of below along with the text of the error for easy copy/pasting
ERROR: Service 'web' failed to build: The command '/bin/sh -c bash INSTALL_NODB.sh -A -u' returned a non-zero code: 1
ERROR MESSAGE SCREENSHOT
Once again, thank you all for any and all help! Please let me know if I can provide any more information!
Looks like this may be an issue that was closed in May of 2021 https://github.com/MISP/MISP/issues/7375. That Docker image has an INSTALL_NODB.sh that was initially committed in March 2021 https://github.com/MISP/misp-docker/commit/1e2f18f2c1211e382bd8df5371b1d3d718dad061. Since it was added before that fix the container may not include the fix for rpi that was added in the main repo. To verify, you can check if the output of uname -m is in this support map from the script used by the docker image https://github.com/MISP/misp-docker/blob/master/web/INSTALL_NODB.sh#L3070. If it isn't, then you would need this fix implemented in the docker image.
aarch64 isn't a supported architecture. There's a pull request on the repository that adds it, so you can add that change to your local repository like this (from a command line in the misp-docker repository):
git remote add fukusuket https://github.com/fukusuket/misp-docker.git
git fetch fukusuket
git merge fukusuket/hotfix/build-error-on-m1-mac -m "add aarch64 support"
Hopefully the pull request will be accepted soon and then you can go back to using the unaltered MISP git repository.

Cannot install any plugin from Jenkins server on setup. Error: Read timeout

I am running Jenkins on docker on my local machine with the official jenkins/jenkins:lts image. Until a few days ago, it still worked like a charm, but now it is unable to install any suggested package.
I checked the logs and found something about unreachable plugin mirror servers.
None of any plugin installed.
Caused: java.net.SocketTimeoutException: Read timed out
Caused: java.io.IOException: Failed to load https://updates.jenkins.io/.../1.7/plain-credentials.hpi to /var/jenkins_home/plugins/plain-credentials.jpi.tmp
I have tried on a couple of machines on different networks and got the same errors, I wonder if it was something relating to server maintenance? I'm from South East Asia. Do you have any idea? Do you have any workaround? Any help is appreciated.
I have been getting similar errors since 2020-11-12. It seems that Jenkins infrastructure is having trouble right now, which is being tracked in the following GitHub issue: https://github.com/jenkins-infra/jenkins.io/issues/3958

Getting a Docker x509 Certificate Error after upgrade

I recently migrated my dev environment from Windows Home to Pro and as part of the upgrade installed Docker Desktop instead of using Docker Toolbox.
After the upgrade, using the default Docker Desktop instance I kept receiving some variation on the below error.
Error response from daemon: Get https://docker.abc.xyz.net/v1/users/:
x509: certificate is valid for *.xyz.net, xyg.net, not
docker.abc.xyz.net
What could be causing the problem? I have searched around the internet for hours and all the different solutions have no effect. Restarting Docker, uninstall and reinstall.
Can anybody help?
It appears as part of the migration from Docker Toolbox to Docker Desktop a number of entries are leftover in the hosts file on Windows that cause a conflict when Docker is trying to access any external network. The simple fix? Remove these 5 lines from my hosts file
52.3.153.154 index.docker.io
34.200.28.105 registry.docker.io
52.20.146.203 registry-1.docker.io
54.152.209.167 auth.docker.io
52.222.149.24 dseasb33srnrn.cloudfront.net

gcloud docker pull fails with Untar exit status 2 unexpected fault address

EDIT: A huge thank you #mattmoor for helping me debug the issue. After I had to create a new docker-machine. There was a problem with the docker daemon that must've arisen due to the first machine not being created correctly.
I am having trouble pulling images from another computer, both of which are running OSX Yosemite. Both machines have the docker daemon running, and have successfully authenticated with the desired project to pull from with
gcloud auth login
On my computer I am able to run:
gcloud docker pull gcr.io/projectid/image-tag without any issues.
However when I try to repeat this on another machine, I get a large error message that begins with:
Error pulling image (tag-here) from gcr.io/projectid/image-tag, endpoint: https://gcr.io/v1/, Untar exit status 2 unexpected fault address 0xc208ce5d04
fatal error: faultr downloading dependent layers
[signal 0xb code=0x1 addr=0xc208ce5d04 pc=0x94109e]
Followed by a goroutine 1 stack trace.
The docker version on both machines is 1.6.2, the client and server api version is 1.18, both Go versions are go1.4.2
The Google Cloud SDK version on both machines is 0.9.67, and both have the following components installed:
bq 2.0.18
bq-nix 2.0.18
core 2015.06.30
core-nix 2015.06.02
gcloud 2015.06.30
gcutil-msg 2015.06.09
gsutil 4.13
gsutil-nix 4.12
preview 2015.06.30
and the machine that works also has these extra components installed:
alpha 2015.06.30
beta 2015.06.30
kubectl
kubectl-darwin-x86_64 0.18.1
Any help would be greatly appreciated, I'm truly baffled as to why I can't pull from the gcr registry on the other machine.
I'm baffled too, this looks like Docker dying trying to untar the blob, and I haven't seen that before.
Would you mind starting a thread with gcr-contact#google.com, as this may take a little debugging, and email will be a bit easier for the back-and-forth.
We can update this with what we find, if that works for you?

Resources