KSAT: Knowledge-infused Self Attention Transformer - Integrating Multiple Domain-Specific Contexts
dc.contributor.author | Roy, Kaushik | |
dc.contributor.author | Zi, Yuxin | |
dc.contributor.author | Narayanan, Vignesh | |
dc.contributor.author | Gaur, Manas | |
dc.contributor.author | Sheth, Amit | |
dc.date.accessioned | 2022-11-03T15:54:12Z | |
dc.date.available | 2022-11-03T15:54:12Z | |
dc.date.issued | 2022 | |
dc.description.abstract | Domain-specific language understanding requires integrating multiple pieces of 2 relevant contextual information. For example, we see both suicide and depression3 related behavior (multiple contexts) in the text “I have a gun and feel pretty bad 4 about my life, and it wouldn’t be the worst thing if I didn’t wake up tomorrow”. 5 Domain specificity in self-attention architectures is handled by fine-tuning on 6 excerpts from relevant domain specific resources (datasets and external knowl7 edge - medical textbook chapters on mental health diagnosis related to suicide 8 and depression). We propose a modified self-attention architecture Knowledge9 infused Self Attention Transformer (KSAT) that achieves the integration of multiple 10 domain-specific contexts through the use of external knowledge sources. KSAT 11 introduces knowledge-guided biases in dedicated self-attention layers for each 12 knowledge source to accomplish this. In addition, KSAT provides mechanics for 13 controlling the trade-off between learning from data and learning from knowledge. 14 Our quantitative and qualitative evaluations show that (1) the KSAT architecture 15 provides novel human-understandable ways to precisely measure and visualize the 16 contributions of the infused domain contexts, and (2) KSAT performs competitively 17 with other knowledge-infused baselines and significantly outperforms baselines 18 that use fine-tuning for domain-specific tasks. | en_US |
dc.description.uri | https://scholarcommons.sc.edu/cgi/viewcontent.cgi?article=1571&context=aii_fac_pub | en_US |
dc.format.extent | 9 pages | en_US |
dc.genre | journal articles | en_US |
dc.genre | preprints | en_US |
dc.identifier | doi:10.13016/m24nqo-eefi | |
dc.identifier.uri | http://hdl.handle.net/11603/26256 | |
dc.language.iso | en_US | en_US |
dc.relation.isAvailableAt | The University of Maryland, Baltimore County (UMBC) | |
dc.relation.ispartof | UMBC Computer Science and Electrical Engineering Department Collection | |
dc.relation.ispartof | UMBC Faculty Collection | |
dc.rights | This item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author. | en_US |
dc.subject | UMBC Ebiquity Research Group | |
dc.title | KSAT: Knowledge-infused Self Attention Transformer - Integrating Multiple Domain-Specific Contexts | en_US |
dc.type | Text | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- KSAT_ Knowledge-infused Self Attention Transformer - Integrating.pdf
- Size:
- 567.17 KB
- Format:
- Adobe Portable Document Format
- Description:
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 2.56 KB
- Format:
- Item-specific license agreed upon to submission
- Description: