Nathan MANZAMBI NDONGALA | Texila American University (original) (raw)

Nathan MANZAMBI NDONGALA

Related Authors

Roshan Chitrakar

Lev Manovich

Lev Manovich

Graduate Center of the City University of New York

Viacheslav Kuleshov

Bogdan Gabrys

PALIMOTE JUSTICE

Nagendra Kumar

Stefan Wermter

Vellingiriraj E.K

Christopher Crick

Uploads

Papers by Nathan MANZAMBI NDONGALA

Research paper thumbnail of Topological Relation Aware Transformer

Texila international journal of academic research, Jan 31, 2024

Research paper thumbnail of Light RAT-SQL: A RAT-SQL with More Abstraction and Less Embedding of Pre-existing Relations

TEXILA INTERNATIONAL JOURNAL OF ACADEMIC RESEARCH

RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding th... more RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding the database relations and questions in a way to improve the semantic parser. In this work, we propose a light version of the RAT-SQL where we dramatically reduced the number of the preexisting relations from 55 to 7 (Light RAT-SQL-7) while preserving the same parsing accuracy. To ensure the effectiveness of our approach, we trained a Light RAT-SQL-2, (with 2 embeddings) to show that there is a statistically significant difference between RAT-SQL and Light RAT-SQL-2 while Light RAT-SQL-7 can compete with RAT-SQL. Keywords: Deep learning, Natural Language Processing, Neural Semantic Parsing, Relation Aware Transformer, RAT-SQL, Text-To-SQL, Transformer.

Research paper thumbnail of Light RAT-SQL: A RAT-SQL with More Abstraction and Less Embedding of Pre-existing Relations

RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding th... more RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding the database relations and questions in a way to improve the semantic parser. In this work, we propose a light version of the RAT-SQL where we dramatically reduced the number of the preexisting relations from 55 to 7 (Light RAT-SQL-7) while preserving the same parsing accuracy. To ensure the effectiveness of our approach, we trained a Light RAT-SQL-2, (with 2 embeddings) to show that there is a statistically significant difference between RAT-SQL and Light RAT-SQL-2 while Light RAT-SQL-7 can compete with RAT-SQL.

Research paper thumbnail of Topological Relation Aware Transformer

Texila international journal of academic research, Jan 31, 2024

Research paper thumbnail of Light RAT-SQL: A RAT-SQL with More Abstraction and Less Embedding of Pre-existing Relations

TEXILA INTERNATIONAL JOURNAL OF ACADEMIC RESEARCH

RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding th... more RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding the database relations and questions in a way to improve the semantic parser. In this work, we propose a light version of the RAT-SQL where we dramatically reduced the number of the preexisting relations from 55 to 7 (Light RAT-SQL-7) while preserving the same parsing accuracy. To ensure the effectiveness of our approach, we trained a Light RAT-SQL-2, (with 2 embeddings) to show that there is a statistically significant difference between RAT-SQL and Light RAT-SQL-2 while Light RAT-SQL-7 can compete with RAT-SQL. Keywords: Deep learning, Natural Language Processing, Neural Semantic Parsing, Relation Aware Transformer, RAT-SQL, Text-To-SQL, Transformer.

Research paper thumbnail of Light RAT-SQL: A RAT-SQL with More Abstraction and Less Embedding of Pre-existing Relations

RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding th... more RAT-SQL is among the popular framework used in the Text-To-SQL challenges for jointly encoding the database relations and questions in a way to improve the semantic parser. In this work, we propose a light version of the RAT-SQL where we dramatically reduced the number of the preexisting relations from 55 to 7 (Light RAT-SQL-7) while preserving the same parsing accuracy. To ensure the effectiveness of our approach, we trained a Light RAT-SQL-2, (with 2 embeddings) to show that there is a statistically significant difference between RAT-SQL and Light RAT-SQL-2 while Light RAT-SQL-7 can compete with RAT-SQL.

Log In