I've got following URL in symfony (specifics not important):
/frontend_dev.php/something/25/apple
... and a routing rule:
/something/:id/:word
The URL works fine when clicked through to on the site, but not when I type in the URL. Instead, symfony says:
Unable to find a matching route to generate url for params "NULL".
The weird thing is that I can navigate to this page and it works, but when hitting Enter in the browser address bar, it no longer finds it.
Any thoughts on what might be the cause of something like this generally?
I should also add that the URL was working fine when typed in the address bar earlier, but doesn't anymore, and I'm not sure what's there that might be interfering with it.
Thanks in advance.
UPDATE:
The exact routing rule:
profile:
url: /profile/:id/:un
param: {module: profile, action: profile}
UPDATE:
Debug toolbar log (the higlighted part):
Unable to find a matching route to generate url for params "NULL". Toggle debug stack
#4 » in sfException::outputStackTrace() from SF_ROOT_DIR\lib\vendor\symfony\lib\exception\sfException.class.php line 110
#3 » in sfException->printStackTrace() from SF_ROOT_DIR\lib\vendor\symfony\lib\controller\sfFrontWebController.class.php line 52
#2 » in sfFrontWebController->dispatch() from SF_ROOT_DIR\lib\vendor\symfony\lib\util\sfContext.class.php line 170
#1 » in sfContext->dispatch() from SF_ROOT_DIR\web\frontend_dev.php line 13
RESOLVED:
Found it....
I had this in my code on the failing page:
<?php echo link_to(__('Back to previous page').' ›', $sf_request->getReferer()) ?>
The referer fails when typing address into bar.
Unless you're neglecting to mention that you navigate to this page usually from forms and it's the lack of POST data that is upsetting the routing; which I'm thinking this is not- this surely must be an issue with your browser.
Have you tried this in a different browser on your system?
If you're still having issues can you copy in the relevant section of your routing.yml?
I was getting this error message on a live site and found the issue to be a link_to command which was being dynamically populated by a field from the database.
However the field was null, so the link_to received no link info.
Related
Any one have solution for this issue .
With Sitecore Solution without SXA Adding empty attribute to httpRunTime tag in web.config file >> requestPathInvalidCharacters="", can fix issue special characters in URL and redirect you to "not found page" insted of 400 error page with message "A potentially dangerous Request.Path value was detected from the client()"
URL like : http://abc/* or http://abc/& >> redirect to item not found page and I replaced it with custom error page . So, NO issue with sitecore site without SXA.
I tried same solution with [sitecore SXA] site but not effected, It go to another error page with message >> illegal characters in path, Also I applied this solution https://kb.sitecore.net/articles/059908, But nothing effect, I still have the same error!!
Any solution, I looking for same behavior in SXA site
I fixed it by using URl rewrite in IIS and created redirect rule to 404 page hosted in site core
I have 1000+ URLs that I want to scrape to retrieve the title value from the HTML. After trying different things, I ultimately used iMacros scripts, which I don't know anything about. Nonetheless, I managed to make a script after reading guides.
My script is working perfectly but has a problem: When leeching URLs titles, if it encounters an HTTP error (e.g. dead link, forbidden page, etc), it crashes with an error message like this one:
Error -1350: Error loading page. Http status 403. Line 4: URL GOTO=http://url.com
Instead of crashing when the script encounters these errors, I would like it to simply skip the URL and continue running. How can I modify my script to do this? Here is my script:
VERSION BUILD=9002379
TAB T=1
TAB CLOSEALLOTHERS
URL GOTO=http://google.com/
ADD !EXTRACT {{!URLCURRENT}}
TAG POS=1 TYPE=TITLE ATTR=* EXTRACT=TXT
SAVEAS TYPE=EXTRACT FOLDER=d:/ FILE=links.txt
SET !EXTRACT_TEST_POPUP NO
Output:
http://google.com/,Google
I would also like to replace the comma after the URL in the output with a semicolon.
At critical points where you don't want to exit at fails:
SET !ERRORIGNORE YES
If you want to revert back at some point to do stop on an error:
SET !ERRORIGNORE NO
You can use these two as many times as you like, even every second row turn on and off.
Validating my feed, it has an enclosure with a URL of
https://archive.org/download/NigelFarageAPersonalMessageToNorthernIrelandVoters./Nigel%20Farage,%20a%20personal%20message%20to%20Northern%20Ireland%20voters..mp3
I know it is a bit convoluted... but what is wrong with it? The stop in the directory name? the double dot in the file name? the comma? all of em?
I have looked at the RFC on URL's but cant make it out(!).
This feed does not validate.
line 441, column 2: url must be a full URL: https://archive.org/download/NigelFarageAPersonalMessageToNorthernIrelandVoters./Nigel%20Farage,%20a%20personal%20message%20to%20Northern%20Ireland%20voters..mp3 (4 occurrences) [help]
<enclosure type="audio/mpeg" url="https://archive.org/download/NigelFarage ...
^
** edit **
A useful (even if incorrect) answer was added (and removed...) showing the result from the w3c URL validator - https://validator.w3.org/checklink
This Link Checker looks for issues in links, anchors and referenced objects in a Web page, CSS style sheet, or recursively on a whole Web site. For best results, it is recommended to first ensure that the documents checked use Valid (X)HTML Markup and CSS. The Link Checker is part of the W3C's validators and Quality Web tools.
If you find this question, you may find the link checker a useful resource!
The problem seems to be that it’s a HTTPS URL instead of a HTTP URL.
The linked error documentation, foo attribute of bar must be a full URL, says:
If this is a link to a web page, you must include the "http://" at the beginning and immediately follow it with a valid domain name.
The RSS 2.0 spec says about <enclosure>:
The url must be an http url.
If you change https://archive.org/download/… to http://archive.org/download/…, it validates.
And if you don't have httpS then your SSL says your page isn't secure. #feedvalidator step up. There are a ton of feedback/complaints about this on the support forum here https://groups.google.com/forum/#!forum/feedvalidator-users
More specifically here: https://github.com/rubys/feedvalidator/issues/16
ISSUE: I'm using FormIt and the redirect hook is not working for me.
It is redirecting to http://www.example.comPageName.html instead of http://www.example.com/PageName.html. It is not placing the forward slash after the page name.
Has anyone seen this issue before?
Here is my FormIt snippet.
[[!FormIt? &hooks=`email,redirect` &emailTpl=`MyEmailChunk` &emailTo=`mypersonalemail#email.com` &emailSubject=`Contact Us form submission` &redirectTo=`35` &emailFrom=`info#domainnamme.com`]]
Yes, the MyEmailChunk exists.
The email addresses are dummy values
The resource with id 35 is in the same context and is published and is hidden from the menus.
I'm using friendly urls.
If I try another resource it works.
URL generated in the following code
$url = $this->modx->makeUrl($this->formit->config['redirectTo'],$contextKey,$redirectParams,'full');
'full' means URL is absolute, prepended with site_url from config ( http://rtfm.modx.com/display/revolution20/modX.makeUrl ). Please check your site_url variable in /core/config/config.inc.php
Actually the site_url variable needs to be checked in the System Settings.
Make sure you have a trailing slash.
In my asp.net mvc 3 site my actual route looks like /FF.mvc/116/MVt?m=01-12-2012 but some of my users are getting error and they have weird route like
/FF.mvc/116/ossw=((qncufuh)niah(`r)mt
any idea where from this
ossw=((qncufuh)niah(`r)mt
coming from?
My hunch is that your applications pages are indexed by the search (google) engine (perhaps against your wishes :)). If you search anything for example your apps name in google you will see a similar ossw=((qncufuh)niah(r)mt` string in the address bar when the results are returned.
Some employee has searched the page link in google and tried to access it from there.
Inorder to prevent the search spiders from indexing your application's pages add a robots.txt file in your application.
Have you tried something like:
Url = "/FF.mvc/116/MVt?m=" + HttpUtility.UrlEncode("01-12-2012")
I don't know if the char '-' will be error.
the ((qncufuh).... might come from some strange language code.