Can look @ TG's Bhunaksha website. It gives more data for survey no. 

http://164.100.187.185/bhunaksha/36/index.jsp


   -  Map Report 
   
<http://164.100.187.185/bhunaksha/36/plotreportTL.jsp?state=36&giscode=B22_1042232008_&plotno=44>
   -  Land Registry Pahani Report 
   
<http://mabhoomi.telangana.gov.in/PahaniViewPPB_bhunaksha.aspx?v=2232008&s=44&T=S&op=B>
   -  Land Registry RoR-1B Report
   -  Dispute Report
   -  Encumbrance Report 
   
<http://registration.telangana.gov.in/ts/DeedDetails.do?method=getDeedDetailsEODB&villageId=2232008&villageName=KHANKHANPET&syNoList=44&syNoTel=44>
   -  Prohibited / Injunction Order Report 
   
<http://registration.telangana.gov.in/ts/UnitRateMV.do?method=getProhbPropertiesWebland&villageCode=2232008&surveyNo=44>

Website for other states are mentioned here. 

http://nlrmp.nic.in/faces/rptPhysicalHome/rptStateGenericDetail.xhtml?id=./../master/physical.xhtml


Thanks,
naveenpf

On Tuesday, 14 November 2017 14:18:29 UTC+5:30, harsha wrote:
>
> Hi Nikhil,
>
> I have been thinking on similar lines to work in Telangana(
> http://mabhoomi.telangana.gov.in/) and have spoken to local land 
> activists & researchers. Why this ? , one is to keep a record dump of 
> records as they are changing very fastly in Telangana with huge amount of 
> surveys done, and we have no clue about how the records are changing, and 
> only the final changes are in public domain. 
> Second we are running a farmer distress helpline since the last 7 months 
> in Vikarabad District, Telangana and 50% of the issues we get are land 
> issues, so it would make the accessibility of the records easy too.
> Third is also to understand and do some analysis on the land acregae, and 
> who owns it, who cultivates/benefits from it(its currently noted in pahani 
> in 13th column). As we have been working on rights of Tenant farmers, this 
> is an important data point and understanding we need to get.
>
> So we would be eager to know on how we can collaborate and take it forward.
> We can take help from Srinivas Kodali([email protected] <javascript:>
> >) , who had locally offerred to help, who has experience in scrapping.
>
> Cheers,
> SreeHarsha
>
>
> On Monday, October 23, 2017 at 2:55:35 PM UTC+5:30, Devendra Damle wrote:
>>
>> Hi Nikhil.
>>
>> A colleague of mine wrote a python script for scraping data from the Debt 
>> Recovery Tribunals website. The problem was similar to yours. 
>>
>> His script uses selenium web driver, and gecko drivers for firefox. It 
>> opens the website in firefox, then simulates clicks to select things from 
>> drop-down menus to generate tables, and then downloads the data in it to a 
>> JSON file. I am attaching the source code file. I am myself not a coder, so 
>> I won't be able to help you with the code itself, but you might be able to 
>> modify it to suit your needs.
>>
>> Regards,
>> Devendra
>>
>> On Monday, October 24, 2016 at 10:09:58 PM UTC+5:30, Nikhil VJ wrote:
>>>
>>> Hi, 
>>>
>>> I'm looking at Maharashtra's land records portal : 
>>> https://mahabhulekh.maharashtra.gov.in 
>>>
>>> .. and wondering if it's possible to scrape data from here? 
>>>
>>> Will share a workflow: 
>>> choose 7/12 (७/१२) > select any जिल्हा > तालुका > गाव         
>>> select शोध :  सर्वे नंबर / गट नंबर (first option) 
>>> type 1 in the text box and press the "शोधा" button 
>>> Then we get a dropdown with options like 1/1 , 1/2, 1/3 etc. 
>>>
>>> On selecting any and clicking "७/१२ पहा", 
>>> a new window/tab opens up (you have to enable popups), having static 
>>> HTML content (some tables). I need to capture this content. 
>>>
>>> The URL is always the same: 
>>> https://mahabhulekh.maharashtra.gov.in/Konkan/pg712.aspx 
>>> ..but the content changes depending on the options chosen. 
>>>
>>> On using the browser's "Inspect Element"> Network and clicking the 
>>> final button, there is a request to this URL: 
>>>
>>> https://mahabhulekh.maharashtra.gov.in/Konkan/Home.aspx/call712 
>>>
>>> and the request Params / Payload is like: 
>>>
>>> {'sno':'1','vid':'273200030398260000','dn':'रत्नागिरी','tn':'खेड','vn':'वाळंजवाडी','tc':'3','dc':'32','did':'32','tid':'3'}
>>>  
>>>
>>>
>>> when you change the survey/gat number to 1/10, the params change like 
>>> so: 
>>> {'sno':'1#10','vid':'273200030398260000','dn':'रत्नागिरी','tn':'खेड','vn':'वाळंजवाडी','tc':'3','dc':'32','did':'32','tid':'3'}
>>>  
>>>
>>>
>>> for 1/1अ: 
>>> {'sno':'1#1अ','vid':'273200030398260000','dn':'रत्नागिरी','tn':'खेड','vn':'वाळंजवाडी','tc':'3','dc':'32','did':'32','tid':'3'}
>>>  
>>>
>>>
>>> I tried some wget and curl commands but no luck so far. Do let me know 
>>> if you can make some headway. 
>>>
>>> Also, it would be great to also learn how to extract on the list of 
>>> districts, talukas (subdistricts) in each district, and villages in 
>>> each taluka. 
>>>
>>> dumping other info at bottom if it helps. 
>>>
>>> Why do this: 
>>> At present it's just an exploration following on from our work on 
>>> village shapefiles. 
>>> The district > taluka > village mapping data from official Land 
>>> Records data could serve as a good source for triangulation. 
>>> Then, while I don't see myself going deeper into this right now, I am 
>>> aware that land records / ownership has major corruption, 
>>> entanglements and other issues precisely because of the lack of 
>>> transparency. The mahabhulekh website itself is a significant step 
>>> forward in making this sector a little more transparent, and more push 
>>> in this direction would probably do more good IMHO. At some point 
>>> GIS/lat-long info might come in, and it would be good to bring the 
>>> data to a level that is ready for it. 
>>>
>>>
>>> Data dump: 
>>> When we press the button to fetch the 7/12 (saatbarah) record, the 
>>> console records a POST with these parameters: 
>>>
>>> Copy as cURL: 
>>> curl 'https://mahabhulekh.maharashtra.gov.in/Konkan/Home.aspx/call712' 
>>> -H 'Host: mahabhulekh.maharashtra.gov.in' -H 'User-Agent: Mozilla/5.0 
>>> (X11; Ubuntu; Linux i686; rv:42.0) Gecko/20100101 Firefox/42.0' -H 
>>> 'Accept: application/json, text/plain, */*' -H 'Accept-Language: 
>>> en-US,en;q=0.5' --compressed -H 'Content-Type: 
>>> application/json;charset=utf-8' -H 'Referer: 
>>> https://mahabhulekh.maharashtra.gov.in/Konkan/Home.aspx' -H 
>>> 'Content-Length: 170' -H 'Cookie: 
>>> ASP.NET_SessionId=3ozsnwd3nhh4py4hmiqcjeoc' -H 'Connection: 
>>> keep-alive' -H 'Pragma: no-cache' -H 'Cache-Control: no-cache' 
>>>
>>> Copy POST data: 
>>> {'sno':'1#1अ','vid':'273200030398260000','dn':'रत्नागिरी','tn':'खेड','vn':'वाळंजवाडी','tc':'3','dc':'32','did':'32','tid':'3'}
>>>  
>>>
>>>
>>> request headers: 
>>> POST /Konkan/Home.aspx/call712 HTTP/1.1 
>>> Host: mahabhulekh.maharashtra.gov.in 
>>> User-Agent: Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:42.0) 
>>> Gecko/20100101 Firefox/42.0 
>>> Accept: application/json, text/plain, */* 
>>> Accept-Language: en-US,en;q=0.5 
>>> Accept-Encoding: gzip, deflate 
>>> Content-Type: application/json;charset=utf-8 
>>> Referer: https://mahabhulekh.maharashtra.gov.in/Konkan/Home.aspx 
>>> Content-Length 
>>> <https://mahabhulekh.maharashtra.gov.in/Konkan/Home.aspxContent-Length>: 
>>> 170 
>>> Cookie: ASP.NET_SessionId=3ozsnwd3nhh4py4hmiqcjeoc 
>>> Connection: keep-alive 
>>> Pragma: no-cache 
>>> Cache-Control: no-cache 
>>>
>>> response headers: 
>>> HTTP/1.1 200 OK 
>>> Cache-Control: private, max-age=0 
>>> Content-Type: application/json; charset=utf-8 
>>> Server: Microsoft-IIS/8.0 
>>> X-Powered-By: ASP.NET 
>>> Date: Mon, 24 Oct 2016 15:31:40 GMT 
>>> Content-Length: 10 
>>>
>>> Copy Response: 
>>> {"d":null} 
>>>
>>>
>>> -- 
>>> -- 
>>> Cheers, 
>>> Nikhil 
>>> +91-966-583-1250 
>>> Pune, India 
>>> Self-designed learner at Swaraj University <
>>> http://www.swarajuniversity.org> 
>>> Blog <http://nikhilsheth.blogspot.in> | Contribute 
>>> <https://www.payumoney.com/webfronts/#/index/NikhilVJ> 
>>>
>>
>>

-- 
Datameet is a community of Data Science enthusiasts in India. Know more about 
us by visiting http://datameet.org
--- 
You received this message because you are subscribed to the Google Groups 
"datameet" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to