Rails Assets intermittent Timeout https://rails-assets.org - ruby-on-rails

Intermittently getting error when doing bundle install for this and also on Continuous Integration server.
Retrying dependency api due to error (4/4): Bundler::HTTPError Network error while fetching https://rails-assets.org/api/v1/dependencies?gems=rails-assets-angular%2Crails-assets-bootstrap-2.3.2%2Crails-assets-bootstrap-3%2Crails-assets-jasmine%2Crails-assets-jqueryjs%2Crails-assets-rainbow%2Crails-assets-typeahead.js (too many connection resets (due to Net::ReadTimeout - Net::ReadTimeout) after 0 requests on 70190348894440, last used 22.186127 seconds ago)

It's likely maintenance happens every once and awhile (especially this late) although they have definitely had downtime before (see https://github.com/tenex/rails-assets/issues/329).
Doing a simple curl you can see they simply aren't responding to any requests:
$ curl 'https://rails-assets.org/api/v1/dependencies?gems=rails-assets-angular%2Crails-assets-bootstrap-2.3.2%2Crails-assets-bootstrap-3%2Crails-assets-jasmine%2Crails-assets-jqueryjs%2Crails-assets-rainbow%2Crails-assets-typeahead.js' -D - --max-time 10
curl: (28) Operation timed out after 10004 milliseconds with 0 bytes received

Related

FatalErrorException Error: Maximum execution time of 30 seconds exceeded

I'm trying to run my symfony 3.3 project in the docker. It's working fine out of docker. All containers are running: nginx:alpine, php-fpm 7.2, PostgreSQL, rabbitMQ.
I'm receiving this error:
(1/1) FatalErrorException
Error: Maximum execution time of 30 seconds exceeded
in AnalyzeServiceReferencesPass.php (line 71)
at AnalyzeServiceReferencesPass->processValue()
in AnalyzeServiceReferencesPass.php (line 118)
dev.log
request.CRITICAL: Uncaught PHP Exception
Symfony\Component\Debug\Exception\FatalErrorException: "Error: Maximum
execution time of 30 seconds exceeded" at
/var/www/html/vendor/twig/twig/lib/Twig/Extension/Core.php line 710
{"exception":"[object]
(Symfony\Component\Debug\Exception\FatalErrorException(code: 0):
Error: Maximum execution time of 30 seconds exceeded at
/var/www/html/vendor/twig/twig/lib/Twig/Extension/Core.php:710)"}
Its means that your request is very slowly, so you need to increase your max_execution time
max_execution_time = 900 or ini_set('max_execution_time', 900);
or change driver of request
also check if you have a infinite cycles
RUN echo "max_execution_time=900" >> /usr/local/etc/php/conf.d/docker-php-ext-xdebug.ini line added to Docker file made it work.

How Can I remove this error "socket: read check timed out(30) sock.c:240: Connection timed out"

I am working on siege 4.0.2 on ubuntu 16.04 environment. I need get a failed transaction when I simulate more than 1100 user, I know that if failed transaction comes means so there is a problem in server memory may be maemory out of failure. How to understand the failed transaction ? And how to solve the problem for failed transaction comes ?
siege -c1190 -t1m http://192.168.1.11:8080/
HTTP/1.1 200 7.02 secs: 57 bytes ==> GET /kiosk/start
HTTP/1.1 200 7.01 secs: 57 bytes ==> GET /kiosk/start
siege aborted due to excessive socket failure; you
can change the failure threshold in $HOME/.siegerc
Transactions: 3263 hits
Availability: 76.11 %
Elapsed time: 9.34 secs
Data transferred: 0.18 MB
Response time: 1.98 secs
Transaction rate: 349.36 trans/sec
Throughput: 0.02 MB/sec
Concurrency: 691.94
Successful transactions: 3263
Failed transactions: 1024
Longest transaction: 7.75
Shortest transaction: 0.03
When I simulate 1100 users , I got a error discriptor tables full sock. c 119: Too many open files and then I do ulimit - n 10000 and the error went.
Then I again simulate 1100 users, I got a new error
[error] socket: read error Connection reset by peer sock.c:539: Connection reset by peer
Could not able to throw the error . How to remove this error ? Anybody please help me

Why are Google Pipeline VM instances hanging indefinitely?

I am using Dockerflow to run parallel tasks through the Google Pipelines API on Google Cloud Platform. I started a single-step task running 1389 VMs in parallel and found that 233 of the VMs were apparently doing nothing and hanging indefinitely.
I did a spot check of the serial console output and repeatedly saw the VMs running into "Getting controller config failed" errors.
When I tried logging into the VMs I received the error: "Connection Failed. We are unable to connect to the VM on port 22".
I am wondering why my VM instances are hanging, and if there is something I can do to avoid running into these issues.
I've included a snippet of the serial console output below
startupscript: +++ readlink -f /usr/share/google-genomics/startup.sh
startupscript: ++ dirname /usr/share/google-genomics/startup.sh
startupscript: + cd /usr/share/google-genomics
startupscript: + ./controller --operation_id <id> --validation_token <token> --base_path https://genomics.googleapis.com
create controller[2905]: Getting controller config
create controller[2905]: Getting controller config failed, will retry: Get <link>: Get <service_account_token_link>: net/http: timeout awaiting response headers
create controller[2905]: Getting controller config failed, will retry: Get <link>: dial tcp 74.125.26.95:443: i/o timeout
collectd[2342]: write_gcm: Asking metadata server for auth token
collectd[2342]: write_gcm: curl_easy_perform() failed: Couldn't connect to server
collectd[2342]: write_gcm: Error -1 from wg_curl_get_or_post
collectd[2342]: write_gcm: wg_transmit_unique_segment failed.
collectd[2342]: write_gcm: wg_transmit_unique_segments failed. Flushing.
there was a temporary networking issue in us-east1-b. All 3 above VMs were in us-east1-b. These minor incidents do not appear in https://status.cloud.google.com/
Serial console output for a successful run looks like:
A Feb 21 19:05:06 ggp-5629907348021283130 startupscript: + ./controller --operation_id --validation_token --base_path https://autopush-genomics.sandbox.googleapis.com
A Feb 21 19:05:06 ggp-5629907348021283130 create controller[2689]: Getting controller config
A Feb 21 19:05:36 ggp-5629907348021283130 create controller[2689]: Getting controller config failed, will retry: Get https://genomics.googleapis.com/v1alpha2/pipelines:getControllerConfig?alt=json&operationId=&validationToken=: dial tcp 173.194.212.81:443: i/o timeout
A Feb 21 19:05:43 ggp-5629907348021283130 controller[2689]: Switching to status: pulling-image
A Feb 21 19:05:43 ggp-5629907348021283130 controller[2689]: Calling SetOperationStatus(pulling-image)
A Feb 21 19:05:44 ggp-5629907348021283130 controller[2689]: SetOperationStatus(pulling-image) succeeded
The "Getting controller config failed, will retry" is fine. It succeeded upon retry. The "SetOperationStatus(pulling-image) succeeded" indicates networking is working.
In theory, you can submit any number of jobs to Pipelines API and the API will take care of queueing.
If these temporary networking hiccups become common, we may consider changing Pipelines API to somehow detect and retry.
there may have been a temporary networking issue. Can you give me some failed operation ids (or failed VM names)?
Have you tried again since then; can you reproduce the problem?

Error installing Crashlytics curl: (35) SSL peer handshake failed, the server most likely requires a client certificate to connect

Installing Crashlytics (3.7.3)
[!] Error installing Crashlytics
[!] /usr/bin/curl -f -L -o /var/folders/2n/hv7xjs3x1mz8988vz8nl3mc40000gn/T/d20160817-1341-a6uhwh/file.zip https://kit-downloads.fabric.io/cocoapods/crashlytics/3.7.3/crashlytics.zip --create-dirs
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0
curl: (35) SSL peer handshake failed, the server most likely requires a client certificate to connect
The previous thread posted 10 days back has been closed by some other user.
Error installing Crashlytics - SSL peer handshake failed
The reply is that it is working fine with other mac osX user. What to do for the current user.

Benchmarking with Siege always returns zero hits and zero failed transactions

All,
I'm learning to do benchmark testing using the Siege tool against our Rails app. I'm running Siege on my OS X box against a website hosted on another server. When I run it I always get zero hits and zero failed transactions no matter which site I run it against.
Because of the limitations of OS X, I've configured more ports with sudo sysctl -w net.inet.tcp.msl=1000 and more open files with launchctl limit maxfiles 10000 10000. I've also configured proxy variables in .siegerc since I'm running behind a proxy. Sample command line is:
siege -c1 -b -t10S 'www.google.com'.
But no matter which website I hit, I always get zero hits, zero failures and zero successful transactions. The Siege log file does not show any errors. What am I doing wrong?
siege -c1 -b -t10S 'www.google.com'
** SIEGE 3.0.5
** Preparing 1 concurrent users for battle.
The server is now under siege...
Lifting the server siege... done.
Transactions: 0 hits
Availability: 0.00 %
Elapsed time: 9.84 secs
Data transferred: 0.00 MB
Response time: 0.00 secs
Transaction rate: 0.00 trans/sec
Throughput: 0.00 MB/sec
Concurrency: 0.00
Successful transactions: 0
Failed transactions: 0
Longest transaction: 0.00
Shortest transaction: 0.00

Resources