CN111353306B - Entity relationship and dependency Tree-LSTM-based combined event extraction method - Google Patents
Entity relationship and dependency Tree-LSTM-based combined event extraction method Download PDFInfo
- Publication number
- CN111353306B CN111353306B CN202010109601.1A CN202010109601A CN111353306B CN 111353306 B CN111353306 B CN 111353306B CN 202010109601 A CN202010109601 A CN 202010109601A CN 111353306 B CN111353306 B CN 111353306B
- Authority
- CN
- China
- Prior art keywords
- vector
- sentence
- entity
- event
- word
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000605 extraction Methods 0.000 title description 18
- 239000013598 vector Substances 0.000 claims abstract description 181
- 238000000034 method Methods 0.000 claims abstract description 21
- 230000002457 bidirectional effect Effects 0.000 claims abstract description 7
- 239000011159 matrix material Substances 0.000 claims description 25
- 238000012549 training Methods 0.000 claims description 8
- 238000002372 labelling Methods 0.000 claims description 7
- 230000001960 triggered effect Effects 0.000 claims description 4
- 238000012935 Averaging Methods 0.000 claims description 3
- 230000011218 segmentation Effects 0.000 claims description 3
- 238000011160 research Methods 0.000 description 3
- 230000007547 defect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/951—Indexing; Web crawling techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Machine Translation (AREA)
Abstract
The invention discloses a method for extracting a combined event based on an entity relationship and a dependency Tree-LSTM. The method comprises the following steps: 1. and encoding the original text and the text marking information. 2. The result of step 1 is input into the bi-directional LSTM. Forward and backward implicit state vectors with timing are obtained. 3. Firstly, an input sentence is analyzed into a dependency Tree structure, then the result in the step 1 is input into the constructed dependency Tree-LSTM, and the Tree root node hidden state vector and the hidden state vector at each moment are obtained. 4. And acquiring and storing the entity relationship sentence information characteristic vector. Connecting forward and backward implicit state vectors of bidirectional LSTM t and implicit state vector depending on Tree-LSTM t time5. Carrying out trigger word identification and classification; 6. event arguments are identified and classified.
Description
Technical Field
The invention relates to an event extraction method, in particular to a combined event extraction method based on an entity relationship and a dependency Tree-LSTM, and belongs to the field of natural language processing.
Background
Event Extraction (EE) is an important component of an Information Extraction task (IE). The Event extraction mainly comprises two subtasks of trigger word recognition and classification (ED) and Event Argument recognition and classification (AI), wherein the ED task mainly finds out the trigger words triggering the events from the text and correctly judges the Event types of the trigger words. The latter is to determine whether the sentence is an event sentence (containing trigger words), and then to determine whether the entity reference appearing in the sentence is the event argument. And assigns each entity mention the correct event argument role. With the occurrence of massive text information and the deep development of deep learning technology, event extraction also becomes a hot problem for research of people. In addition, the event extraction technology has been applied to news message classification, social public opinion management, and the like.
Disclosure of Invention
The invention provides a combined event extraction method based on an entity relationship and a dependency Tree-LSTM, which is mainly aimed at the problems that the dependency path of an event trigger word and an event argument is too long and the output characteristics of a model lack the entity relationship.
The method for extracting the combined event based on the entity relationship and the dependency Tree-LSTM comprises the following steps:
step 1, encoding an original text and text labeling information;
step 2, inputting the result of the step 1 into a bidirectional LSTM; obtaining forward implicit state vectors with timingAnd backward implicit state vector
Step 3, firstly, analyzing an input sentence into a dependency Tree structure by using a Stanford CoreNLP tool, then inputting the coding result of the step 1 into a dependency Tree-LSTM constructed by the dependency Tree structure, and obtaining a Tree root node hidden state vectorAnd implicit State vectors at t instants
Step 4, carrying out entity relation vector RkEncoding a junction tree root node hidden state vectorObtaining and saving entity relation sentence vectorForward implicit state vectors connecting simultaneously two-way LSTM t instantsAnd backward implicit state vectorAnd an implicit State vector that depends on the Tree-LSTM t timeSolving new implicit state vectorsThereby not only saving the information of the sub-nodes but also acquiring the local context information with a certain time sequence;
step 5, connecting the hidden state vector H at the time t in the step 4tCarrying out trigger word recognition and classification with the sentence vector F;
further, the step 1 is specifically realized as follows:
1-1, acquiring unprocessed original text and text labeling information from a source file, wherein the labeling information comprises entity mention, entity types, event trigger words, event arguments, event argument roles, entity relationships and entity relationship argument roles, wherein the number of the entity types is 7, the number of the event trigger word types is 39, the number of the entity relationship types is 20, and the number of the entity relationship argument roles is 16; then, sentence and word segmentation are carried out on the original text by using Stanford CoreNLP; acquiring a dependency tree structure of parts of speech and sentences, wherein each word is used as a node of the tree structure; respectively creating a part-of-speech vector table, an entity type vector table, an entity relation argument role vector table, a trigger word type vector table and an event argument role vector table, wherein each vector table has initialization vectors corresponding to other types; entity mentions may consist of multiple words; for convenience of representing entity mentions, we denote each entity mention by the head of each entity mention (mostly the last word of an entity mention) and the subscript that the head appears in sentences denotes the subscript of each entity mention; thus, the subscripts referred to by each entity are indicated by the symbols: head1,head2,head3,...,headk-1,headk(where k is the number of entity mentions, k may be zero); for this purpose, we useRepresenting entity mentions that appear in the sentence; randomly initializing each vector in all vector tables, and updating the vectors during training;
1-2, inquiring the pre-trained glove word vector matrix to obtain the word vector w of each word in the sentenceiThen, the part-of-speech vector table is inquired to obtain a part-of-speech vector wposAnd querying the entity type vector table to obtain an entity type vector we;
Obtain each word representation xi={wi,wpos,weThus the sentence vector matrix is denoted W ═ x1,x2,...,xn-1,xnWhere n is the length of the sentence;
further, step 2 is specifically implemented as follows:
converting the vector matrix W of the sentence into { x }1,x2,...,xn-1,xnInputting the sentence into a bidirectional LSTM, and respectively acquiring a forward implicit state matrix of the sentenceAnd backward implicit state matrixWhereinAndrepresenting the forward hidden state vector and the backward hidden state at time t, respectively, t ∈ [1, n]Bi-directional LSTM is a time series sensitive model and, therefore,andrespectively storing the upper and lower information with certain time sequence information;
further, step 3 is specifically implemented as follows:
analyzing each sentence into a tree structure through a Stanford CoreNLP tool, wherein each word in the sentence forms a node of the tree structure, and the parent node or the child node of the node appears when the dependency relationship exists between the word and the node; changing W to { x ═ x1,x2,...,xn-1,xnInputting the information into a dependency Tree-LSTM constructed based on the Tree structure, and obtaining an implicit state vector of each node in the Tree structure parsed by the sentenceAnd the implicit state vector of the root nodeThus, the dependency Tree-LSTM of a sentence outputs the implicit state matrix of the sentenceWherein t, root ∈ [1, n ]]N is the length of the sentence;
further, step 4 is specifically implemented as follows:
4-1, obtaining an entity relation vector R in the sentence by inquiring the entity relation table initialized randomly in the step 1kRepresenting the kth entity relationship; if no entity relationship exists, RkPoint to the "other" entity relationship vectors and adjust the vectors during the training process;
the memory unit vector c and the hidden state vector h of each node in the 4-2 dependency Tree-LSTM are obtained by summing the hidden state vectors of the sub-nodes of the node; therefore, the root node in the semantic dependency tree structure contains the information of the whole sentence, and the hidden vector of the root node generated in the step 4 is used for making the sentence contain the sentence-level vector of the entity relationship informationAnd an entity relationship vector RkConnecting and obtaining sentence vector containing entity relation information
4-3, combining the hidden vectors at each moment in the step 2 and the step 3, and simultaneously, acquiring the hidden state vector at the moment t by adopting an averaging mode to reduce the dimensionality of the hidden vectors:and the hidden state matrix of the whole sentence is H ═ H1,H2,···,Hn-1,HnWhere t ∈ [1, n ]]N is the length of the sentence;
further, step 5 is specifically implemented as follows:
5-1 specifies onlyThere are verbs and nouns as trigger word candidates, and there are 39 seed types in total, including "other" types; judging the part of speech of each word in the sentence, if the part of speech is a verb or a noun, carrying out hidden state vector H at the current t momenttThe expression is connected with the sentence vector F and is input into a trigger word multi-classification formula:
wherein, WTAnd bTRespectively triggering a weight matrix and a bias item of multi-classification of words;representing the probability that the trigger word candidate for the tth word (each word being a time instant) triggers the event type,representing the event type triggered at the t-th moment;
further, step 6 is specifically implemented as follows:
6-1, 20 entity relation argument roles are provided, a randomly initialized entity relation argument role vector table is established, the vector table is searched through the entity relation argument roles, and the vector is adjusted in the training process; by usingRepresenting the ith entity mention in an entity relationship vector RkPlays the role of j entity relationship argument;
6-2, referring the entity in the sentence as an event argument candidate word; sequentially subjecting the ith event argument candidate word (the ith entity reference) to implicit state vectorTth recognized as a trigger word at step 5-1Implicit state vector H of a wordtSentence vector F containing entity relationship and ith event argument candidate in entity relationship RkEntity relationship argument role inConnecting; inputting the join vector into an event argument recognition multi-classification formula:
wherein, WAAnd bARespectively, a weight matrix and bias terms for the event argument class,indicating the event type of the ith event argument candidateProbability values of the role of event argument played;indicating the event type of the ith event argument candidateThe role of event argument;
the invention has the following beneficial effects:
aiming at the defects of the prior art, a method for extracting the combined event based on the entity relationship and the dependency Tree-LSTM is provided. And obtaining the hidden state vector of each moment by using a dependency Tree-LSTM and a bidirectional LSTM, and respectively combining the entity relationship vector and the entity relationship argument role vector with the hidden state vectors to perform multi-classification on the trigger word candidate words and the argument candidate words. The model not only can reduce the influence of wrong trigger word types on argument identification, but also can fully utilize entity relationship and entity relationship argument role information, thereby improving the accuracy of the event extraction model.
Drawings
FIG. 1 is a flow chart of the overall implementation of the present invention.
FIG. 2 is a detailed flow diagram of the present invention triggering word recognition and classification and event argument recognition and classification.
Fig. 3 is a network architecture diagram of the model of the invention.
Detailed Description
The attached drawings disclose a flow chart of a preferred embodiment of the invention in a non-limiting way; the technical solution of the present invention will be described in detail below with reference to the accompanying drawings.
The event extraction is an important component of information extraction research and is a common technical basis for news hotspot extraction and social public opinion analysis. The event extraction is to find out event suggestions from a large amount of texts, and the event suggestions are composed of event trigger words and event arguments. Therefore, the event extraction mainly comprises two tasks of trigger word recognition and event argument role classification. Some researches divide the task into two stages, wherein the first stage firstly acquires the event type of the trigger word, and then judges the role of the event argument candidate word in the sentence according to the category of the trigger word. The method has the defect that the error classification of the trigger words in the first stage influences the effect of event argument role classification, so that a joint learning model of trigger word identification and event argument classification is provided. However, the above model does not take full advantage of the entity relationships and the role of entity references in entity relationship arguments. Therefore, a method for extracting the combined event based on the entity relationship and the dependency Tree-LSTM is provided.
1-3, a method for extracting join events based on entity relationships and dependency Tree-LSTM, comprises the following steps:
step 1, encoding the original text and the text label information.
Step 2 inputs the result of step 1 into the bi-directional LSTM. Obtaining forward implicit state vectors with timingAnd backward implicit state vector
Step 3, firstly, the Stanford CoreNLP tool is utilized to analyze the input sentence into a dependency Tree structure, then the result of the step 1 is input into a dependency Tree-LSTM constructed by the dependency Tree structure, and the hidden state vector of the Tree root node is obtainedAnd an implicit state vector for each time instant
Step 4, entity relation RkCoded connectionObtaining and storing entity relation sentence information characteristic vectorAt the same time, the forward implicit state vector of the concatenated bi-directional LSTM tAnd backward implicit state vectorAnd an implicit State vector that depends on the Tree-LSTM t timeMake itThe information of the sub-nodes can be saved, and the local context information with a certain time sequence can be obtained.
Step 5, connecting the hidden state vector H at the time t in the step 4tCarrying out trigger word recognition and classification with the sentence vector F;
Further, the step 1 is specifically realized as follows:
the method comprises the steps of obtaining unprocessed original texts and labeling information from a source file, wherein the labeling information comprises entity words, entity types, event trigger words, event arguments, event argument roles, entity relationships and entity relationship argument roles, and the labeling information comprises 7 entity types, 39 event trigger word types, 20 entity relationship types and 16 entity relationship argument roles. And then, sentence and word segmentation are carried out on the original text by using the Stanford CoreNLP. And acquiring a dependency tree structure of parts of speech and sentences, wherein each word is used as a node of the tree structure. And respectively creating a part-of-speech vector table, an entity type vector table, an entity relation argument role vector table, a trigger part-of-speech type vector table and an event argument role vector table, wherein each vector table has other corresponding initialization vectors. These vectors are initialized randomly and updated at the time of training.
Inquiring the pre-trained glove word vector matrix to obtain the word vector w of each word in the sentenceiThen, the part-of-speech vector table is inquired to obtain wposAnd querying entity type to obtain we。
Each word obtained is represented by xi={wi,wpos,weThus the sentence vector matrix is denoted W ═ x1,x2,...,xn-1,xnWhere n is the length of the sentence.
Converting the vector matrix W of the sentence into { x }1,x2,...,xn-1,xnInputting the sentence into a bidirectional LSTM, and respectively acquiring a forward implicit state matrix of the sentenceAnd backward implicit state matrixWhereinAndrepresenting the forward hidden state vector and the backward hidden state at time t, respectively, t ∈ [1, n]Bi-directional LSTM is a time series sensitive model and, therefore,andthe above and below information with certain timing information is saved separately.
The Stanford CoreNLP tool parses each sentence into a tree structure, with each word in the sentence constituting a node of the tree structure, where a dependency relationship with the word occurs with either a parent or child of the node. Changing W to { x ═ x1,x2,...,xn-1,xnInputting the information into the dependency Tree-LSTM constructed based on the Tree structure, and obtaining the implicit state vector of each node in the Tree structure analyzed by the sentenceAnd the implicit state vector of the root nodeDependency Tree-LSTM of a sentence thus outputs an implicit state matrix of the sentenceWherein t, root ∈ [1, n ]]And n is the length of the sentence.
In event extraction, some trigger words may be ambiguous in recognition, for example: elop plan to leveNokia. Most event extraction models (EE) identify leave as an event type transport more easily, but if the relationship of an entity Elop in a sentence and membership existing in an entity Nokia is utilized, the EE can identify an End-Position event triggered by leave in the sentence more easily. Therefore, by inquiring the entity relationship table initialized randomly in the step (1), the entity relationship vector R in the sentence is obtainedk(indicating a kth entity relationship), and if no entity relationship exists, RkPoint to the "other" entity relationship vector and adjust the vector during the training process.
The memory element vector c and the hidden state vector h of each node in the dependency Tree-LSTM are obtained by summing the hidden state vectors of the sub-nodes of the node. Therefore, the root node in the semantic dependency tree structure contains the information of the whole sentence, and the hidden vector of the root node generated in step 4 is used for making the sentence contain the sentence-level vector of the entity relationship informationAnd an entity relationship vector RkConnecting and obtaining sentence vector containing entity relation information
The dependency Tree-LSTM is a non-time-series sensitive model, and the implicit state vector output at each time also lacks certain time-series information, so the implicit vectors at each time in steps 2 and 3 are combined, but in order to reduce the dimensionality of the implicit vectors, the implicit state vector at time t is obtained by averaging:and the hidden state matrix of the whole sentence is H ═ H1,H2,…,Hn-1,HnWhere t ∈ [1, n ]]N is a sentenceLength.
Specify that only verbs and nouns are candidates for trigger words, for a total of 39 seed types, including "other" types. Firstly, judging the part of speech of each word in a sentence, and if the part of speech is a verb or a noun, carrying out hidden state vector H at the current t momenttThe expression is connected with the sentence vector F and is input into a trigger word multi-classification formula:
wherein,the probability that the trigger word candidate for the tth word triggers the event type,indicating the event type triggered by the t-th word.
For judging the event argument role played by the event argument candidate word (entity mention) in the sentence in the event type, it is desirable to utilize the entity relation argument role played by the entity mention in the entity relation. As with the example sentence mentioned in 4-1, the two entity mentions Elop and Nokia, if learned by the model, act as employeemberber and org, respectively, in the entity relationship membership. The model will more easily assign event argument roles Person and Entity to the two event arguments Elop and Nokia in the event type transport. The total number of the entity relationship argument roles is 20, a randomly initialized entity relationship argument role vector table is established, the table is searched through the entity relationship argument roles, and the vectors are adjusted in the training process. By usingEntity mention at time i in entity relationship RkPlays the role of j entity relationship argument.
In the sentenceEntity mentions as event argument candidates. Sequentially subjecting ith event argument candidate word to implicit state vector HiImplicit State vector connection H for the t-th word identified as trigger word at 5-1tSentence vector F containing entity relationship and ith event argument candidate in relationship RkEntity relationship argument role inAnd (4) connecting. Inputting the join vector into an event argument recognition multi-classification formula:
Claims (7)
1. The method for extracting the combined event based on the entity relationship and the dependency Tree-LSTM is characterized by comprising the following steps of:
step 1, encoding an original text and text labeling information;
step 2, inputting the result of the step 1 into a bidirectional LSTM; obtaining forward implicit state vectors with timingAnd backward implicit state vector
Step 3, firstly, analyzing an input sentence into a dependency Tree structure by using a Stanford CoreNLP tool, then inputting the coding result of the step 1 into a dependency Tree-LSTM constructed by the dependency Tree structure, and obtaining a Tree root node hidden state vectorAnd implicit State vectors at t instants
Step 4, carrying out entity relation vector RkEncoding a junction tree root node hidden state vectorObtaining and saving entity relation sentence vectorForward implicit state vectors connecting simultaneously two-way LSTM t instantsAnd backward implicit state vectorAnd an implicit State vector that depends on the Tree-LSTM t timeSolving new implicit state vectorsThereby both preserving sub-nodesThe information also acquires local context information with a certain time sequence;
step 5, connecting the hidden state vector H at the time t in the step 4tCarrying out trigger word recognition and classification with the sentence vector F;
step 6, sequentially identifying the hidden state vector H of the t-th word identified as the trigger word in the step 5tThe ith event argument candidate word, namely the ith entity mention hidden state vectorSentence vector F containing entity relationship and entity relationship vector R of ith event argument candidatekEntity relationship argument role inAnd connecting, and identifying and classifying event arguments.
2. The method for extracting combined event based on entity relationship and dependency Tree-LSTM as claimed in claim 1, wherein step 1 is implemented as follows:
1-1, acquiring unprocessed original text and text label information from a source file, wherein the label information comprises entity mention, entity type, event trigger word, event argument role, entity relationship and entity relationship argument role, wherein the label information comprises 7 entity types, 39 event trigger word types, 20 entity relationship types and 16 entity relationship argument roles; then, sentence and word segmentation are carried out on the original text by using Stanford CoreNLP; acquiring a dependency tree structure of parts of speech and sentences, wherein each word is used as a node of the tree structure; respectively creating a part-of-speech vector table, an entity type vector table, an entity relation argument role vector table, a trigger word type vector table and an event argument role vector table, wherein each vector table has initialization vectors corresponding to other types;
1-2, inquiring a pre-trained glove word vector matrix to obtain a word vector w of each word in a sentenceiThen, the part-of-speech vector table is inquired to obtain the wordProperty vector wposAnd querying the entity type vector table to obtain an entity type vector we;
Obtain each word representation xi={wi,wpos,weThus the sentence vector matrix is denoted W ═ x1,x2,...,xn-1,xnWhere n is the length of the sentence.
3. The method for extracting combined event based on entity relationship and dependency Tree-LSTM as claimed in claim 1 or 2, wherein step 2 is implemented as follows:
converting the vector matrix W of the sentence into { x }1,x2,...,xn-1,xnInputting the sentence into a bidirectional LSTM, and respectively acquiring a forward implicit state matrix of the sentenceAnd backward implicit state matrixWhereinAndrepresenting the forward hidden state vector and the backward hidden state at time t, respectively, t ∈ [1, n]Bi-directional LSTM is a time series sensitive model and, therefore,andthe above and below information with certain timing information is saved separately.
4. The method for extracting combined events based on entity relationship and dependency Tree-LSTM as claimed in claim 3, wherein step 3 is implemented as follows:
analyzing each sentence into a tree structure through a Stanford CoreNLP tool, wherein each word in the sentence forms a node of the tree structure, and the parent node or the child node of the node appears when the dependency relationship exists between the word and the node; changing W to { x ═ x1,x2,...,xn-1,xnInputting the information into a dependency Tree-LSTM constructed based on the Tree structure, and obtaining an implicit state vector of each node in the Tree structure parsed by the sentenceAnd the implicit state vector of the root nodeThus, the dependency Tree-LSTM of a sentence outputs the implicit state matrix of the sentenceWherein t, root ∈ [1, n ]]And n is the length of the sentence.
5. The method for extracting combined events based on entity relationship and dependency Tree-LSTM as claimed in claim 4, wherein step 4 is implemented as follows:
4-1, obtaining an entity relation vector R in the sentence by inquiring the entity relation table initialized randomly in the step 1kRepresenting the kth entity relationship; if no entity relationship exists, RkPoint to the "other" entity relationship vectors and adjust the vectors during the training process;
the memory unit vector c and the hidden state vector h of each node in the 4-2 dependency Tree-LSTM are obtained by summing the hidden state vectors of the sub-nodes of the node; therefore, the root node in the semantic dependency tree structure contains the information of the whole sentence, and the hidden vector of the root node generated in the step 4 is used for making the sentence contain the sentence-level vector of the entity relationship informationAnd entitiesRelation vector RkConnecting and obtaining sentence vector containing entity relation information
4-3, combining the hidden vectors at each moment in the step 2 and the step 3, and simultaneously, acquiring the hidden state vector at the moment t by adopting an averaging mode to reduce the dimensionality of the hidden vectors:and the hidden state matrix of the whole sentence is H ═ H1,H2,···,Hn-1,HnWhere t ∈ [1, n ]]And n is the length of the sentence.
6. The method for extracting combined events based on entity relationship and dependency Tree-LSTM as claimed in claim 5, wherein step 5 is implemented as follows:
5-1, only verbs and nouns are specified as trigger word candidates, and 39 types of seeds are provided in total, wherein the types comprise 'other' types; judging the part of speech of each word in the sentence, if the part of speech is a verb or a noun, carrying out hidden state vector H at the current t momenttThe expression is connected with the sentence vector F and is input into a trigger word multi-classification formula:
Pt tri=softmaxtri(WT[Ht,F]+bT)
7. The method for extracting combined event based on entity relationship and dependency Tree-LSTM as claimed in claim 6, wherein step 6 is implemented as follows:
6-1, 20 entity relation argument roles are provided, a randomly initialized entity relation argument role vector table is established, the vector table is searched through the entity relation argument roles, and the vector is adjusted in the training process; by usingRepresenting the ith entity mention in an entity relationship vector RkPlays the role of j entity relationship argument;
6-2, mentioning the entity in the sentence as a candidate word of event argument; sequentially subjecting ith event argument candidate word hidden state vectorImplicit State vector H for the tth word identified as the trigger word in step 5-1tSentence vector F containing entity relationship and ith event argument candidate in entity relationship RkEntity relationship argument role inConnecting; inputting the join vector into an event argument recognition multi-classification formula:
wherein, WAAnd bARespectively, a weight matrix and bias terms for the event argument class,indicating the ith event argument candidate in the event classModel (III)Probability values of the role of event argument played;indicating the event type of the ith event argument candidatePlays the role of event argument.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010109601.1A CN111353306B (en) | 2020-02-22 | 2020-02-22 | Entity relationship and dependency Tree-LSTM-based combined event extraction method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010109601.1A CN111353306B (en) | 2020-02-22 | 2020-02-22 | Entity relationship and dependency Tree-LSTM-based combined event extraction method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111353306A CN111353306A (en) | 2020-06-30 |
CN111353306B true CN111353306B (en) | 2020-10-16 |
Family
ID=71195780
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010109601.1A Active CN111353306B (en) | 2020-02-22 | 2020-02-22 | Entity relationship and dependency Tree-LSTM-based combined event extraction method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111353306B (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112417878B (en) * | 2020-11-24 | 2024-06-14 | 北京明略昭辉科技有限公司 | Entity relation extraction method, system, electronic equipment and storage medium |
CN112541364A (en) * | 2020-12-03 | 2021-03-23 | 昆明理工大学 | Chinese-transcendental neural machine translation method fusing multilevel language feature knowledge |
CN112507077B (en) * | 2020-12-15 | 2022-05-20 | 杭州电子科技大学 | Event time sequence relation identification method based on relational graph attention neural network |
CN112559713B (en) * | 2020-12-24 | 2023-12-01 | 北京百度网讯科技有限公司 | Text relevance judging method and device, model, electronic equipment and readable medium |
CN112784576B (en) * | 2021-01-13 | 2022-07-29 | 哈尔滨工程大学 | Text dependency syntactic analysis method |
CN113158667B (en) * | 2021-04-09 | 2022-03-01 | 杭州电子科技大学 | Event detection method based on entity relationship level attention mechanism |
CN115794444B (en) * | 2023-02-02 | 2023-05-16 | 广州钛动科技股份有限公司 | Event communication method, event communication device, computer equipment and computer readable storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102693219A (en) * | 2012-06-05 | 2012-09-26 | 苏州大学 | Method and system for extracting Chinese event |
CN105260361A (en) * | 2015-10-28 | 2016-01-20 | 南京邮电大学 | Trigger word tagging system and method for biomedical events |
CN107992476A (en) * | 2017-11-28 | 2018-05-04 | 苏州大学 | Towards the language material library generating method and system of Sentence-level biological contexts network abstraction |
CN110598001A (en) * | 2019-08-05 | 2019-12-20 | 平安科技(深圳)有限公司 | Method, device and storage medium for extracting association entity relationship |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104809105B (en) * | 2015-05-11 | 2017-12-26 | 苏州大学 | Recognition methods and the system of event argument and argument roles based on maximum entropy |
US20170357625A1 (en) * | 2016-06-14 | 2017-12-14 | Northrop Grumman Systems Corporation | Event extraction from documents |
EP3407209A1 (en) * | 2017-05-22 | 2018-11-28 | Fujitsu Limited | Apparatus and method for extracting and storing events from a plurality of heterogeneous sources |
US11416817B2 (en) * | 2017-06-02 | 2022-08-16 | Apple Inc. | Event extraction systems and methods |
CN108875809A (en) * | 2018-06-01 | 2018-11-23 | 大连理工大学 | The biomedical entity relationship classification method of joint attention mechanism and neural network |
CN109635280A (en) * | 2018-11-22 | 2019-04-16 | 园宝科技(武汉)有限公司 | A kind of event extraction method based on mark |
CN109657041B (en) * | 2018-12-04 | 2023-09-29 | 南京理工大学 | Deep learning-based automatic problem generation method |
-
2020
- 2020-02-22 CN CN202010109601.1A patent/CN111353306B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102693219A (en) * | 2012-06-05 | 2012-09-26 | 苏州大学 | Method and system for extracting Chinese event |
CN105260361A (en) * | 2015-10-28 | 2016-01-20 | 南京邮电大学 | Trigger word tagging system and method for biomedical events |
CN107992476A (en) * | 2017-11-28 | 2018-05-04 | 苏州大学 | Towards the language material library generating method and system of Sentence-level biological contexts network abstraction |
CN110598001A (en) * | 2019-08-05 | 2019-12-20 | 平安科技(深圳)有限公司 | Method, device and storage medium for extracting association entity relationship |
Non-Patent Citations (1)
Title |
---|
基于依存树的越南语新闻事件元素抽取技术研究;周晶晶;《中国优秀硕士学位论文全文数据库 信息科技辑》;20170215(第02期);I138-4499页 * |
Also Published As
Publication number | Publication date |
---|---|
CN111353306A (en) | 2020-06-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111353306B (en) | Entity relationship and dependency Tree-LSTM-based combined event extraction method | |
CN114020768B (en) | Method for constructing SQL sentence generation model of Chinese natural language and application | |
CN106599032B (en) | Text event extraction method combining sparse coding and structure sensing machine | |
US9672205B2 (en) | Methods and systems related to information extraction | |
Demir et al. | Improving named entity recognition for morphologically rich languages using word embeddings | |
CN105988990B (en) | Chinese zero-reference resolution device and method, model training method and storage medium | |
Viola et al. | Learning to extract information from semi-structured text using a discriminative context free grammar | |
US5991710A (en) | Statistical translation system with features based on phrases or groups of words | |
CN109145260B (en) | Automatic text information extraction method | |
CN111241294A (en) | Graph convolution network relation extraction method based on dependency analysis and key words | |
Fonseca et al. | A two-step convolutional neural network approach for semantic role labeling | |
Mori et al. | A machine learning approach to recipe text processing | |
CN111274829B (en) | Sequence labeling method utilizing cross-language information | |
CN116628173B (en) | Intelligent customer service information generation system and method based on keyword extraction | |
CN107818141A (en) | Incorporate the biomedical event extraction method of structuring key element identification | |
CN113821605B (en) | Event extraction method | |
US7627567B2 (en) | Segmentation of strings into structured records | |
CN107526721B (en) | Ambiguity elimination method and device for comment vocabularies of e-commerce products | |
CN111091009A (en) | Document association auditing method based on semantic analysis | |
CN111160027A (en) | Cyclic neural network event time sequence relation identification method based on semantic attention | |
CN112966525A (en) | Law field event extraction method based on pre-training model and convolutional neural network algorithm | |
Stoeckel et al. | Voting for POS tagging of Latin texts: Using the flair of FLAIR to better ensemble classifiers by example of Latin | |
KR101072460B1 (en) | Method for korean morphological analysis | |
CN113158667B (en) | Event detection method based on entity relationship level attention mechanism | |
CN114881043A (en) | Deep learning model-based legal document semantic similarity evaluation method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
EE01 | Entry into force of recordation of patent licensing contract |
Application publication date: 20200630 Assignee: Hangzhou Yuanchuan New Technology Co.,Ltd. Assignor: HANGZHOU DIANZI University Contract record no.: X2021330000781 Denomination of invention: Joint event extraction method based on entity relationship and dependent tree LSTM Granted publication date: 20201016 License type: Common License Record date: 20211206 |
|
EE01 | Entry into force of recordation of patent licensing contract |