import.io

Data crawling using import.io [closed]

﹥>﹥吖頭↗ 提交于 2020-01-06 07:05:30
问题 Closed . This question needs to be more focused. It is not currently accepting answers. Want to improve this question? Update the question so it focuses on one problem only by editing this post. Closed 4 years ago . i have crawled some data from import.io. That data are in my import.io account. I have another import.io account. Now I want to tranfer data of my previous account to my new account. How can i do it? Is it even possible? I found no option for that Thanks in advance 回答1: Do you

import.io curl simple data integration

可紊 提交于 2020-01-03 05:31:09
问题 I want to get the simple api integration from https://import.io via curl. I created a magic api and struggle now to get the data. Do I use the magic api on the browser, it works fine. But with curl I get a {"code":"NOT_AUTHORIZED"} . Obvious that the authentication did not work. Here's my code #!/bin/bash login=foo@bar.com pass=foobar userid=foo-bar apikey=1f2o3o4b5a6r curl --user $login:$pass https://api.import.io/store/data/033fbe66-13b7-4cf2-ba9c-58183a567a6f/_query?input/webpage/url=https

Extract value from javascript object in site using xpath and import.io

我们两清 提交于 2019-12-31 01:51:37
问题 I want extract a number provided by javascript object in site, but I really don't understand that I am doing. I tried different versions using alike examples and guidelines in import.io site and other tutorial sites, but I got only 1 of two results: extracted all numbers on given page or nothing at all. I tried e.g. //[contains(.,"Unikālo apmeklējumu skaits:")]@type ; //[contains(.,"Unikālo apmeklējumu skaits:")] . Most likely it's necessary to add there something else, but I just don't know

How to send crawler data to PHP via command line?

杀马特。学长 韩版系。学妹 提交于 2019-12-23 02:42:29
问题 Can I send the results rather than stored in the JSON file, send it to PHP? I have this two files settings.json { "outputFile" : "C:\\wamp\\www\\drestip\\admin\\crawls\\mimshoes.json", "logFile" : "C:\\wamp\\www\\drestip\\admin\\crawls\\mimshoes.tsv", "pause" : 1, "local" : false, "connections" : 3, "cookiesEnabled" : false, "robotsDisabled" : false, "advancedMode" : true, "crawlTemplate" : [ "www.mimshoes.com/" ], "startUrls" : [ PAGES ], "maxDepth" : 10, "dataTemplate" : [ "www.mimshoes.com

How to use substring() with Import.io?

守給你的承諾、 提交于 2019-12-22 06:56:21
问题 I'm having some issues with XPath and import.io and I hope you'll be able to help me. :) The html code: <a href="page.php?var=12345"> For the moment, I manage to extract the content of the href ( page.php?var=12345 ) with this: ./td[3]/a[1]/@href Though, I would like to just collect: 12345 substring might be the solution but it does not seem to work on import.io as I use it... substring(./td[3]/a[1]/@href,13) Any ideas of what the problem is? Thank's a lot in advance! 回答1: Try using this for

Running a exe file (import.io) through PHP in Ubuntu

拈花ヽ惹草 提交于 2019-12-12 06:05:30
问题 I need to call an installed tool - Import.io using PHP in Ubuntu. Whenever I click on the PHP button it should navigate to import.io tool installed in the system and open import.io. how can I do it? 回答1: Files ending in .exe are Windows binaries, and won't run on Ubuntu natively. You might be able to get it to run on Wine, a Windows emulator for Linux. However, since this is import.io, I suspect it would be much easier, and much more reliable, to call their API instead. You can do this with

basic import.io html search

帅比萌擦擦* 提交于 2019-12-11 03:40:21
问题 So if any of you have any experience with scraping or particularly import.io it will help, since import.io is what i'm using... although I think my question is just about JS really... I really just want to connect a basic html input to the import.io JS code so I can have a custom search http://jsfiddle.net/LSng3/1/ "input": { var search_name = document.getElementsByName("search_name").value; "search_name": search_name } <input name="search_name" placeholder="doesnt work :("> Heres my go...

How to send crawler data to PHP via command line?

半城伤御伤魂 提交于 2019-12-08 19:02:27
Can I send the results rather than stored in the JSON file, send it to PHP? I have this two files settings.json { "outputFile" : "C:\\wamp\\www\\drestip\\admin\\crawls\\mimshoes.json", "logFile" : "C:\\wamp\\www\\drestip\\admin\\crawls\\mimshoes.tsv", "pause" : 1, "local" : false, "connections" : 3, "cookiesEnabled" : false, "robotsDisabled" : false, "advancedMode" : true, "crawlTemplate" : [ "www.mimshoes.com/" ], "startUrls" : [ PAGES ], "maxDepth" : 10, "dataTemplate" : [ "www.mimshoes.com/{alpha}-{alpha}_{alpha}-{alpha}$" ], "destination" : "JSON", "connectorGuid" :

Data crawler or something else

旧街凉风 提交于 2019-12-08 04:27:45
问题 I'm looking for something that I don't know exactly how it can be done. I don't have deep knowledge into crawling, scrapping and etc, but I believe the kind of technology I'm looking for are these. I've a list of around 100 websites that I'd like to monitor constantly. At least once every 3 or 4 days. In these website's I'd look for some logical matches, like: Text contains 'ABC' AND doesn't contain 'BCZ" OR text contains 'XYZ' AND doesn't contain 'ATM' and so on so forth The tool would have

import.io JSON data

大兔子大兔子 提交于 2019-12-06 07:36:31
问题 I need to output some values from a JSON generated by import.io This is the JSON: object(stdClass)#1 (7) { ["offset"]=> int(0) ["results"]=> array(5) { [0]=> object(stdClass)#2 (3) { ["vanzare"]=> string(6) "4.4400" ["moneda"]=> string(3) "EUR" ["cumparare"]=> string(6) "4.3550" } [1]=> object(stdClass)#3 (3) { ["vanzare"]=> string(6) "4.1882" ["moneda"]=> string(3) "USD" ["cumparare"]=> string(6) "4.0582" } [2]=> object(stdClass)#4 (3) { ["vanzare"]=> string(6) "6.1327" ["moneda"]=> string(3