toplogo
Log på

Replication Study of BASS Framework for Abstractive Summarization


Kernekoncepter
The authors conducted a detailed replication study of the BASS framework, focusing on challenges in replicating key components and discrepancies in performance compared to the original work.
Resumé
The study replicates the BASS framework for abstractive summarization based on Unified Semantic Graphs. Challenges in replication, discrepancies in performance, and recommendations for writing replicable papers are highlighted. The study emphasizes the importance of clear technical descriptions and self-explanatory details to ensure successful replication. The authors implemented components like pre-processing, graph construction, text encoder alignment, and model architecture. Discrepancies between replicated methods and original paper were identified. Challenges with missing information, algorithmic complexity, and error-proneness were encountered during replication. Key findings include differences between USGsrc and USGppr structures, challenges in aligning graph and text embeddings, and issues with graph propagation using PageRank. Recommendations focus on providing clear technical context, notation precision, and commented pseudo-code for better reproducibility.
Statistik
The BigPatent dataset includes 1,204,631 training documents. The BASS model had 201M parameters. Training models took about 4 hours on average. Pre-processing runtime was up to 10 hours per chunk. Model ended up having approximately 205M trainable parameters.
Citater
"We found some inconsistencies between author information, paper details, and source code." "Our results indicate poor performance due to model architecture rather than graph structure." "Challenges included missing third-party information and ambiguity of self-explanatory details."

Dybere Forespørgsler

How can missing third-party information impact the replicability of research studies?

Missing third-party information, such as version details and configurations of external components like libraries or tools, can significantly impact the replicability of research studies. Without this crucial information, replicating the study becomes challenging as it may lead to discrepancies in results due to differences in versions or settings. Researchers attempting replication may face obstacles in accurately reproducing the experimental setup, leading to potential errors or inaccuracies in their findings. Additionally, missing third-party information can hinder the understanding of key processes or methodologies used in the original study, further complicating replication efforts.

What are the implications of discrepancies between replicated methods and original papers?

Discrepancies between replicated methods and original papers have several implications for research reproducibility. Firstly, these discrepancies can cast doubt on the validity and reliability of both the original study and its replication. It raises concerns about whether the results reported in either work are accurate and trustworthy. Moreover, inconsistencies between replicated methods and original papers make it difficult for researchers to build upon existing findings or implement proposed techniques effectively. This lack of consistency hampers scientific progress by introducing uncertainty into subsequent studies that rely on previous research.

How can algorithmic complexity affect the accuracy of replication studies?

Algorithmic complexity plays a significant role in determining the accuracy of replication studies. Complex algorithms increase the likelihood of errors during implementation, especially if detailed descriptions or clear instructions are lacking from the original paper. Replicators may struggle with interpreting intricate algorithms correctly, leading to deviations from intended processes and potentially incorrect outcomes. The more complex an algorithm is, the higher the chance for misinterpretation or misimplementation during replication efforts. Simplifying algorithms where possible or providing comprehensive explanations can help mitigate issues related to algorithmic complexity when conducting replication studies.
0
visual_icon
generate_icon
translate_icon
scholar_search_icon
star