You can inspect the page. By either pressing CTRL + SHIFT + I or CMD + OPTION + I on mac, or you can just double click and click inspect.
You click on the “elements” tab and then press CTRL + F or CMD + F, then you paste the like you wanna check.
Check this example.
Hope it helped
Solution 2 :
If there is a link in a web page, it should mostly be in an <a> tag within the href attribute. So, it is a matter of opening the console and getting all the <a> tags and collecting their href.
var allLinkElements = document.querySelectorAll("a");
var allLinks= []
for (var i=0; i<allLinkElements .length; i++){
allLinks.push(allLinkElements [i].href);
}
console.log(allLinks.includes('http://www.what-im-looking-for.com'));
I know I can use following Google search: site:https://example.com text to find
although it only works for texts.
Any idea if it’s possible to find links in href that or other way?
Comments
Comment posted by ahrefs.com/backlink-checker
It sounds like you are wanting to search for “backlinks” (links that point back to your site)… if so, and this is a one time thing there are lots of backlink checker tools out there e.g. here’s one:
Comment posted by What topics can I ask about here?
Iām voting to close this question because it is not a programming question.
Comment posted by engray
@scunliffe thanks for the tool! This is exactly what I want to do, although not for checking backlinks linking to my website, but I want to be sure some links I had in WordPress page are not hardcoded in WP template. I have WP admin access, but no FTP rights. To be precise my question is about finding such links via google search
Comment posted by engray
@Rob my question is about Google search, page you’ve linked contains point “software tools commonly used by programmers; and is”. Is Google not a tool commonly used by programmers? š
Comment posted by engray
Thanks for reply, I know I could do it this way, but what I’m looking for is finding all hrefs in all sites in some domain. I’ve edited my question – now I read it and it’s really not too specific š
Comment posted by Tomas Mota
You could try using beautiful soup with python maybe. That would let you get all of the hrefs within a page.
Comment posted by engray
Thanks for reply, I know I could do it this way, but what I’m looking for is finding all hrefs in all sites in some domain. I’ve edited my question – now I read it and it’s really not too specific š
Comment posted by Syed
look its not specific thats correct. it is universally understood that you need source code in database to extract term aka ‘href’ from the each source code file. how you do it is up to you, grab the sitemap file and fetch cource code of each link. could use any simple old php or phython and run regex to extract terms into array. No one will hand it for free its what the big business is all about.
Comment posted by stackoverflow.com/questions/1439326/ā¦
here is a previous answer on the subject that might help