stanford-nlp

Extract Relationship Between two Entities using StanfordCoreNLP

独自空忆成欢 提交于 2019-12-08 03:24:11
问题 the similar question has been asked here but i cant find any relevant answer of it so i am trying again. I am able to get NER and Dependency tree using the library. Now what i looking for is that i want to extract entities with the relationship between the entities. For example , "flipkart has invested in myntra" so i should be able to get entity1 as "flipkart " and entity2 as "myntra" and "investor" as the relation .or similar kind of structure. I am not able to get correct tool for doing

Is it possible to get a set of a specific named entity tokens that comprise a phrase

纵饮孤独 提交于 2019-12-07 23:44:29
I'm using the Stanford CoreNLP parsers to run through some text and there are date phrases, such as 'the second Monday in October' and 'the past year'. The library will appropriately tag each token as a DATE named entity, but is there a way to programmatically get this whole date phrase? And it's not just dates, ORGANIZATION named entities will do the same ("The International Olympic Committee", for example, could be one identified in a given text example). String content = "Thanksgiving, or Thanksgiving Day (Canadian French: Jour de" + " l'Action de grâce), occurring on the second Monday in

stanford dependency parser

廉价感情. 提交于 2019-12-07 22:34:20
问题 I have tried stanford dependency parser. I got the following parse tree and relation.But I need a dependency graph. How to get it. Is there any way to covert the dependencies to graph? Please help me. I am new to java and stanford tools. Program is a set of instruction (ROOT (S (NP (NNP Program)) (VP (VBZ is) (NP (NP (DT a) (NN set)) (PP (IN of) (NP (NN instruction))))))) nsubj(set-4, Program-1) cop(set-4, is-2) det(set-4, a-3) root(ROOT-0, set-4) prep_of(set-4, instruction-6) 回答1: You must

Extract Relationship Between two Entities using StanfordCoreNLP

别等时光非礼了梦想. 提交于 2019-12-07 17:42:28
the similar question has been asked here but i cant find any relevant answer of it so i am trying again. I am able to get NER and Dependency tree using the library. Now what i looking for is that i want to extract entities with the relationship between the entities. For example , "flipkart has invested in myntra" so i should be able to get entity1 as "flipkart " and entity2 as "myntra" and "investor" as the relation .or similar kind of structure. I am not able to get correct tool for doing the same. I neee some guidelines guys, how to achieve this?. Thanks in advance You can train your own

Stanford NLP - OpenIE out of memory when processing list of files

十年热恋 提交于 2019-12-07 14:42:15
问题 I'm trying to extract information from several files using the OpenIE tool from Stanford CoreNLP, it gives an out of memory error when several files are passed to the input, instead of just one. All files have been queued; awaiting termination... java.lang.OutOfMemoryError: GC overhead limit exceeded at edu.stanford.nlp.graph.DirectedMultiGraph.outgoingEdgeIterator(DirectedMultiGraph.java:508) at edu.stanford.nlp.semgraph.SemanticGraph.outgoingEdgeIterator(SemanticGraph.java:165) at edu

Stanford Core NLP: Entity type non deterministic

做~自己de王妃 提交于 2019-12-07 11:43:33
问题 I had built a java parser using Stanford Core NLP. I am finding an issue in getting the consistent results with the CORENLP object. I am getting the different entity types for the same input text. It seems like a bug to me in CoreNLP. Wondering if any of the StanfordNLP users have encountered this issue and found workaround for the same. This is my Service class which I am instantiating and reusing. class StanfordNLPService { //private static final Logger logger = LogConfiguration.getInstance

Creating Spark schema for GLoVe word vector files

半腔热情 提交于 2019-12-07 11:26:08
问题 GLoVe pre-trained word vectors which can be downloaded here (https://nlp.stanford.edu/projects/glove/) have the following file format: government 0.38797 -1.0825 0.45025 -0.23341 0.086307 -0.25721 -0.18281 -0.10037 -0.50099 -0.58361 -0.052635 -0.14224 0.0090217 -0.38308 0.18503 0.42444 0.10611 -0.1487 1.0801 0.065757 0.64552 0.1908 -0.14561 -0.87237 -0.35568 -2.435 0.28428 -0.33436 -0.56139 0.91404 4.0129 0.072234 -1.2478 -0.36592 -0.50236 0.011731 -0.27409 -0.50842 -0.2584 -0.096172 -0.67109

How to extract Named Entity + Verb from text

非 Y 不嫁゛ 提交于 2019-12-07 10:58:30
问题 Well, my aim is to extract NE (Person) and a verb connected to it from a text. For example, I have this text: Dumbledore turned and walked back down the street. Harry Potter rolled over inside his blankets without waking up. As an ideal result i should get Dumbledore turned walked; Harry Potter rolled I use Stanford NER to find and mark persons, then I delete all sentences that don't contain NE. So, in the end I have a 'pure' text that consists only of sentences with names of characters.

Stanford CoreNLP Error creating edu.stanford.nlp.time.TimeExpressionExtractorImpl

丶灬走出姿态 提交于 2019-12-07 08:18:56
问题 I am trying to learn the Stanford CoreNLP library. I am using C# with the posted example (https://sergeytihon.wordpress.com/2013/10/26/stanford-corenlp-is-available-on-nuget-for-fc-devs/). I loaded the package “Stanford.NLP.CoreNLP” (it added IKVM.NET) via nuget and downloaded the code. Unzipped the .jar models. My directory is correct. I get the following error: > edu.stanford.nlp.util.ReflectionLoading.ReflectionLoadingException was > unhandled HResult=-2146233088 Message=Error creating >

Spanish POS tagging with Stanford NLP - is it possible to get the person/number/gender?

隐身守侯 提交于 2019-12-07 02:23:26
I'm using Stanford NLP to do POS tagging for Spanish texts. I can get a POS Tag for each word but I notice that I am only given the first four sections of the Ancora tag and it's missing the last three sections for person, number and gender. Why does Stanford NLP only use a reduced version of the Ancora tag? Is it possible to get the entire tag using Stanford NLP? Here is my code (please excuse the jruby...): props = java.util.Properties.new() props.put("tokenize.language", "es") props.put("annotators", "tokenize, ssplit, pos, lemma, ner, parse") props.put("ner.model", "edu/stanford/nlp/models