From 29ac3d5c3fcb64b7bfba3b601f663438da3bb21e Mon Sep 17 00:00:00 2001 From: Matthias Veigel Date: Sun, 18 May 2025 13:24:24 +0200 Subject: [PATCH] Added selection criteria and review process overview --- main.typ | 68 ++++++++++--- review_process.drawio | 231 ++++++++++++++++++++++++++++++++++++++++++ review_process.svg | 3 + 3 files changed, 290 insertions(+), 12 deletions(-) create mode 100644 review_process.drawio create mode 100644 review_process.svg diff --git a/main.typ b/main.typ index 0461ca8..4a7efa8 100644 --- a/main.typ +++ b/main.typ @@ -45,6 +45,19 @@ todo todo = Methodology +This publication was created following the process described in @process_fig. The protocol for the review is divided up into the object of the research see @research_questions_s, the search strategy see @sas_s, the selection criteria see @selection_criteria_s and the data extraction strategy see @data_extraction_s. +#place( + bottom + center, + scope: "parent", + float: true, + [ + #figure( + caption: [Overview of the review process], + image("review_process.svg") + ) + ] +) + == Objective and research questions The goal of this research paper is to find claims about the advantages and disadvantages of using dataflow analysis for compiler optimization and where DFA is already implemented in Compilers. This goal has been defined in two research questions: @@ -54,9 +67,8 @@ This goal has been defined in two research questions: - RQ2 --- How is dataflow analysis used in current compilers? \ This questions aims to identify how DFA is already used in current compilers and if it is used during normal compilation or if it has to be explicitly enabled. -== Search and selection strategy +== Search and selection strategy My search strategy consisted of 4 steps as seen in @sas_fig. \ - #figure( caption: [Search string used in electronic databases], kind: "raw", @@ -69,7 +81,6 @@ My search strategy consisted of 4 steps as seen in @sas_fig. \ ``` ] ) - The papers from the first steps are collected from the electronic databases ACM Digital Library, IEEE Xplore, Springer Link, Web of Science with the search string seen in @sas_search_string. The search string in @sas_search_string was created using the research questions in @research_questions_s and was always applied to the full text of the papers. \ In the second step all duplicates which where returned from multiple databases where removed from the results. \ @@ -77,17 +88,19 @@ In the third step the selection was filtered by applying all selection criteria In the forth step I snowballed the previously acquired results. This was to find relevant papers which where not included because of either the search string or the search criteria. \ Afterwards all papers where evaluated based on the data extraction items mentioned in @data_extraction_s. #place( - bottom + center, + auto, scope: "parent", float: true, [ #set par(leading: 0.3em) - #set text(size: 9pt) + #set text(size: 8pt) #figure( caption: [Search and selection process], cetz.canvas({ import cetz.draw: * - let bs = (3.3, 1.1) + let bs = (2.8, 1) + + set-style(stroke: (thickness: 0.5pt)) rect((0, 0), (rel: bs), name: "acm") rect((0, -(bs.at(1)+0.3)*1), (rel: bs), name: "ieee") @@ -104,7 +117,7 @@ Afterwards all papers where evaluated based on the data extraction items mention line("websci.east", (rel: (0.75, 0)), name: "dld") line("dlu.end", "dld.end", name: "dl") - set-style(mark: (end: ">")) + set-style(mark: (end: "straight")) line("dl.50%", "dup.west") line("dup.east", "sel.west") line("sel.east", "snow.west") @@ -117,23 +130,47 @@ Afterwards all papers where evaluated based on the data extraction items mention content("dup", align(center)[Duplicate removal \ n = ]) content("sel", align(center)[Application of \ selection criteria \ n = ]) content("snow", align(center)[Snowballing \ n = ]) - content("inc", align(center)[Papers included \ n = ]) + content("inc", align(center)[Publications included \ n = ]) }) ) ] ) == Selection criteria +For a publication to be relevant it has to satisfy at least one inclusion criteria and not any exclusion criteria. The criteria were chosen to include as any publications as possible but still filter out irrelevant ones. #[ + #v(10pt) #set enum(numbering: (.., i) => "IC" + str(i)) + + Publications comparing DFA to other optimization techniques + + Publications focusing on a specific compiler (e.g., LLVM, Java JIT, C\# JIT) + + Publications providing an implementation for a DFA optimization + #v(10pt) ] +_IC1_ was chosen to help answer _RQ1_. \ +_IC2_ is to include publications which talk about a compiler and how DFA is implemented in it. \ +_IC3_ is to further include publications which directly provide an implementation #[ + #v(10pt) #set enum(numbering: (.., i) => "EC" + str(i)) + + Publications which discuss DFA in a non-compiler context + + Publications written in a language other than english + + Secondary and tertiary publications (e.g., systematic literaturer reviews, surveys) + + Publications in the form of tutorial papers, short papers, poster papers, editorials + + Publications for which the full text is not available + #v(10pt) ] +_EC1_ is to exclude publications which talk about DFA in other contexts which are not relevant to compiler optimization. \ +_EC2-EC5_ are to exclude publications which don't provide enough information to include them in this publication. == Data extraction +Based on the research questions I wrote 9 data items to exrtract from all included publications. @data_extraction_table lists all data items. \ +Data items _D1-D3_ are to document the source of the publication. \ +_D4_ and _D5_ are to explicitly list the advantages and disadvantages for answering _RQ1_. \ +_D6_ and _D7_ show in which compiler DFA was implemented and if it is running directly on a programming language like C++ or if it runs on a intermediate language like LLVM IR. \ +_D8_ lists which optimizations where performed based on the results of DFA and _D9_ lists the limitations of the executed DFA. (e.g., only run on function scope). \ +All data items were extracted from the full text of all included publications. #place( - bottom + center, + auto, scope: "parent", float: true, [ @@ -154,19 +191,26 @@ Afterwards all papers where evaluated based on the data extraction items mention ([Title], [Documentation]), ([Named advantage(s) of DFA for CO], [RQ1]), ([Named disadvantage(s) of DFA for CO], [RQ1]), - ([Analyzed compilers], [RQ2]), - ([In what way is DFA used], [RQ2]) + ([Analyzed compiler(s)], [RQ2]), + ([Targeted language(s) of the optimization], [RQ2]), + ([What optimizations are implemented with DFA], [RQ2]), + ([Limitations of the analysis], [RQ2]) ).enumerate(start: 1).map(((i, arr)) => ([D#i], ..arr)).flatten() ) ) ] ) -#bibliography("refs.bib", title: "References", style: "association-for-computing-machinery") +#colbreak() +#set heading(numbering: none) += References +#bibliography("refs.bib", title: none, style: "association-for-computing-machinery") +/* #colbreak(weak: true) #set heading(numbering: "A.a.a") = Artifact Appendix In this section we show how to reproduce our findings. +*/ diff --git a/review_process.drawio b/review_process.drawio new file mode 100644 index 0000000..a2d0415 --- /dev/null +++ b/review_process.drawio @@ -0,0 +1,231 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/review_process.svg b/review_process.svg new file mode 100644 index 0000000..5d21abe --- /dev/null +++ b/review_process.svg @@ -0,0 +1,3 @@ + + +
Planning
Planning
Research
questions
Research...
Protocol
Protocol
Review need
identification
Review need...
Research questions
definition
Research questions...
Protocol
definition
Protocol...
Protocol
evaluation
Protocol...
Conducting
Conducting
Database
search
Database...
Selection
review
Selection...
Snowballing
Snowballing
Data extraction
Data extraction
Data
synthesis
Data...
Set of
publications
Set of...
Relevant publications
Relevant pub...
Data
items
Data...
Reporting
Reporting
Final report
Final report
Threats analysis
Threats analysis
Report writing
Report writing
Report evaluation
Report evaluation
Activity
Activity
Artifact
Artifact
Activity Flow
Activity Flow
Artifact Flow
Artifact Flow
\ No newline at end of file