-
Notifications
You must be signed in to change notification settings - Fork 9
Feature Request: Curl Page Title #71
Comments
Okay this gonna need some time but it's possible 👍 But wouldn't it be better to catch the http code instead of the page title @mitchellkrogza ? |
Yes possibly easier to decifer between |
Clear !! |
👍 Awesome looking forward to giving that a test run. |
@funilrys Check this cheap and nasty .csv file output showing domain, status code, content type and redirect url (if any). With this .csv file, one can take anything with a Then one can run a separate test on the redirect url's column to see what those produce 😁 and then merge them together and run a test using a list of domains that did not redirect + the ones that did. See: The script .... as I said .... cheap and nasty but .... fast and effective. Took Travis CI 26 minutes to produce the .csv on a list of 5451 domains 👍 |
@funilrys sent you something on WeTransfer 👍 😀 |
Let me check I was out with friends .... |
Okay for this issue the hardest thing is to develop some procedure/features and have to think about how data will be shown in a HTML ( #62 ) file which I started to design 😆 about WeTransfer: Answered you:+1: :open_mouth: |
Looking forward to seeing what you busy developing. Just came up with that
basic bash script this morning. No doubt you will improve upon it greatly
:)
|
The following may be me merged to the wiki after implementation of this issue and if we share the same thought about those codes For your code list, you should consider the following codes @mitchellkrogza : As active
As potentially active
As inactive or potentially inactive
|
Cool thanks Nissar, still lots of playing to do with this one. Trying to
add page title into the equation then it will also help diagnosing lists.
Kind Regards
Mitchell Krog
**************************************************
Visit me at https://mitchellkrog.com
**************************************************
License My Images From Getty Images Here
<http:https://www.gettyimages.com/search/photographer?family=creative&page=1&photographer=mitchell%20krog&sort=best&excludenudity=true#license>
or From Gallo Images Here
<http:https://galloimages.co.za/Search?q=mitchell%20krog&p=1&a=1&l=2,1&st=2&dr=on&is=1&token=48034&pp=13&rc=6>
**************************************************
|
@mitchellkrogza Updated my last comments about the codes For the "checking redirection" part imagine that the redirection have a redirection which also have a redirection 😜 🤣 🤣 |
@funilrys yes indeed those with multiple redirects are the one's I am most interested in. I test all stuff added to my lists manually in a browser before they get added. I always run screen recorder to capture what's happening in the url bar of the browser as I often test a site and it does 1-7 redirects in a split second. So then I play back the screen recording and capture all those redirect links and then add them to my lists. ....... very time consuming as you can imagine 😬 |
Are you suggesting that I should add a follow redirection for funceble once we have the curl column ? 🤣 🤣 🤣 🤣 🤣 |
It would probably kill funceble and Travis too 🤣 ..... this should be a separate project, a redirect-redirect checker. |
Indeed yeah 🤣 imagine a dead-hosts with follow redirection 🤣 🤣 🤣 🤣 |
That could go awfully wrong very fast 🤣 🤣 🤣 |
YEAH .... now we are heading in the right direction |
Another teasing 😉
|
Nice 👍 question, will running funceble on my repo like https://github.com/mitchellkrogza/Stop.Google.Analytics.Ghost.Spam.HOWTO automatically create the new folders if needed? and populate them with a Looking forward to what's coming 😁 |
Thank you for the question @mitchellkrogza !! Please report to #89 😉 👍 |
Perhaps doing a simple curl to get the page title of a page would help people sort and clean their lists. This could be include in results.txt or even separate output files??? So pages that return a name other than 404 or not found go into one result file and ones that return a page name other than 404, missing or not found go into a separate file. This way, especially with lists that have subdomains of a root domain like blogspot.com etc it would be easy to pick up blogs that no longer exist.
The reason for this is that lists that are full of sub-domains of a main domain, like this https://github.com/funilrys/dead-hosts/blob/master/add.2o7Net/tested-list/add.2o7Net.list first of all always show ACTIVE in results because the root domain is active. Getting a page title for each tested domain / sub-domain would quickly reveal which of those dub-domains are no longer actually there. Same applies to some lists that have hundreds of .blogspot.com domains in them.
Something like this?
The text was updated successfully, but these errors were encountered: