bigdata

Which one is best: Apache Ambari cluster on Physical system with 5 Machine or install on virtual machine with diffrent 5 VM?

戏子无情 提交于 2020-01-07 04:55:10
问题 Hi I am working on One of my project which I have created VM of 5 Machine and it is working fine in development environment but I have some confusion regarding VM cluster is good or need to go with physical system cluster. 回答1: Hadoop was developed for physical systems but it will function with varying degrees of success in virtual environments, it depends on the specific environment. This is actually quite a common question on the hadoop mailing lists and was specifically addressed by the

Computing cosine similarities on a large corpus in R using quanteda

断了今生、忘了曾经 提交于 2020-01-07 03:04:47
问题 I am trying to work with a very large corpus of about 85,000 tweets that I'm trying to compare to dialog from television commercials. However, due to the size of my corpus, I am unable to process the cosine similarity measure without getting the "Error: cannot allocate vector of size n" message, (26 GB in my case). I am already running R 64 bit on a server with lots of memory. I've also tried using the AWS on the server with the most memory, (244 GB), but to no avail, (same error). Is there a

Trying to install HUE but no success

我们两清 提交于 2020-01-06 19:47:10
问题 I am trying to install hue on ubuntu and gets the following error when trying to install. Can anyone please tell me why it's giving error for lber.h? I have installed all of the dependencies and using hue2.1.0 Thanks 回答1: What is your version of Ubuntu? Hue works well with the LTS 12:04 and 14:04. Also make sure that you have installed the specific LDAP packages: https://github.com/cloudera/hue#development-prerequisites 来源: https://stackoverflow.com/questions/26143050/trying-to-install-hue

Apache Spark - Scala - HashMap (k, HashMap[String, Double](v1, v2,..)) to ((k,v1),(k,v2),…)

自古美人都是妖i 提交于 2020-01-06 16:18:51
问题 I got: val vector: RDD[(String, HashMap[String,Double])] = [("a", {("x",1.0),("y", 2.0),...}] I want to get: RDD[String,(String,Double)] = [("a",("x",1.0)), ("a", ("y", 2.0)), ...] How can it be done with FlatMap? Better solutions are welcome! 回答1: Try: vector.flatMapValues(_.toSeq) 来源: https://stackoverflow.com/questions/38507249/apache-spark-scala-hashmap-k-hashmapstring-doublev1-v2-to-k-v1

Optimizing a very huge mysql table (query or mysql)

微笑、不失礼 提交于 2020-01-06 12:50:24
问题 I have a mysql DB with 50 GB data and 200M record in a table. I am running following query and it takes 350 second to complete: SELECT x_date, count(*) as totl FROM `rec_ex_15` WHERE x_date > '2014-12-01' and typx = '2' group by x_date order by x_date desc x_date and typx are indexed. Here is the explain: id select_type table type possible_keys key key_len ref rows 1 SIMPLE rec_ex_15 range typx,x_date x_date 3 NULL 15896931 Using where Are there any way to get the result faster? 回答1: As noted

Nifi ExecuteGroovyScript - class already loaded in another classloader

吃可爱长大的小学妹 提交于 2020-01-06 10:15:20
问题 I get a flowfile with ExecuteGroovyScript processor with some custom code in it. and it work well : but if I stop it and change the code in i get this error: java.lang.UnsatisfiedLinkError: Native Library /data/nifi_flow/dec-enr/pseudo/lib/libpseudojni.so already loaded in another classloader: java.lang.UnsatisfiedLinkError: Native Library /data/nifi_flow/dec-enr/pseudo/lib/libpseudojni.so already loaded in another classloader java.lang.UnsatisfiedLinkError: Native Library /data/nifi_flow/dec

How to predict correctly in sklearn RandomForestRegressor?

天大地大妈咪最大 提交于 2020-01-06 04:54:06
问题 I'm working on a big data project for my school project. My dataset looks like this: https://github.com/gindeleo/climate/blob/master/GlobalTemperatures.csv I'm trying to predict the next values of "LandAverageTemperature". First, I've imported the csv into pandas and made it DataFrame named "df1". After taking errors on my first tries in sklearn, I converted the "dt" column into datetime64 from string then added a column named "year" that shows only the years in the date values.-Its probably

Java handling billions bytes

两盒软妹~` 提交于 2020-01-06 02:26:17
问题 I'm creating a compression algorithm in Java ; to use my algorithm I require a lot of information about the structure of the target file. After collecting the data, I need to reread the file. <- But I don't want to. While rereading the file, I make it a good target for compression by 'converting' the data of the file to a rather peculiar format. Then I compress it. The problems now are: I don't want to open a new FileInputStream for rereading the file. I don't want to save the converted file

updating line in large text file using scala

拟墨画扇 提交于 2020-01-05 12:32:36
问题 i've a large text file around 43GB in .ttl contains triples in the form : <http://www.wikidata.org/entity/Q1001> <http://www.w3.org/2002/07/owl#sameAs> <http://la.dbpedia.org/resource/Mahatma_Gandhi> . <http://www.wikidata.org/entity/Q1001> <http://www.w3.org/2002/07/owl#sameAs> <http://lad.dbpedia.org/resource/Mohandas_Gandhi> . and i want to find the fastest way to update a specific line inside the file without rewriting all next text. either by updating it or deleting it and appending it

updating line in large text file using scala

老子叫甜甜 提交于 2020-01-05 12:32:26
问题 i've a large text file around 43GB in .ttl contains triples in the form : <http://www.wikidata.org/entity/Q1001> <http://www.w3.org/2002/07/owl#sameAs> <http://la.dbpedia.org/resource/Mahatma_Gandhi> . <http://www.wikidata.org/entity/Q1001> <http://www.w3.org/2002/07/owl#sameAs> <http://lad.dbpedia.org/resource/Mohandas_Gandhi> . and i want to find the fastest way to update a specific line inside the file without rewriting all next text. either by updating it or deleting it and appending it